Search Results

Search found 47408 results on 1897 pages for 'database machine'.

Page 495/1897 | < Previous Page | 491 492 493 494 495 496 497 498 499 500 501 502  | Next Page >

  • How is network mounted software executed?

    - by CptSupermrkt
    I would like to understand how network mounted software works. For example, at my place of work, we have a software server. Each client machine (hundreds of them) automatically mounts directories from the software server on boot. For example, a program like Matlab is installed just once on the software server, but each client machine can start up an instance of Matlab. What is going on under the hood? Let's say I run /opt/bin/matlab and /opt/ is mounted from the software server, what happens when I press Enter to execute matlab on a client machine? The process is on the client machine, and I've already narrowed down that there isn't any implicit or hidden file transfer (i.e. copying matlab to my machine temporarily for that session) by running matlab on a computer with nearly zero disk space (i.e. not enough room to transfer). Since Matlab was installed on the server, how is my client computer executing it? What mechanism is controlling this? What is happening behind the scenes?

    Read the article

  • How do I start the Workstation Service so I can use `net use`?

    - by nitefrog
    I have a Windows 7 machine that logs into a domain. The machine can net view and see the different shares, but when I try to use, net use * \\name\share, I get an error stating that the service is not started. Yet when I issue a net start, it states the service is already started. My other win7 machines work fine; they can see and use any of the shares. Is there a security setting that needs to be disabled or enabled? I really need to get this working, and I have no other ideas as the other machines have no problem accessing the shares on different systems. The error I am getting is , "The Workstation Service Has Not Been Started", but like I said other machines can connect fine, and when I issue a, "net start workstation", it states the service is already started. In addition the error number I am receiving is 2138. UPDATE: On the machine that is having issues. From the troubled machine if I issue a Net View \\name I can see all the shares on the machine I want to connect to. When I try to net use * \\name\sharename I get the error The Workstation service has not started. I have set both settings sc config lanmanworkstation start = auto and sc config lanmanserver start = auto on the Windows7 computer that is having issues. I have rebooted the computer and still no dice. I can net view any computer on the network and see all shares, but I cannot access any of the shares in which I can see. In the registry under the HKLM\System\CurrentControlSet\Services Both LanmanServer and LanmanWorkstation start is set to 2. Screen capture of net use and view: The Services: This is really weird. What am I missing? It has to be a security setting...

    Read the article

  • NIC is receiving, but not transmitting at all?

    - by Shtééf
    I'm trying to fix a very strange problem remotely on a machine at a customer site. The machine is a Dell PowerEdge, I believe a 1950 (haven't verified, but the lspci output matches specs I found.) The machine has two similar NICs, identified as Broadcom Corporation NetXtreme II BCM5708 Gigabit Ethernet (rev 12) by lspci, and using the bnx2 driver. (I suspect these are on-board and on the same controller, which is what I'm accustomed to for this type of machine.) The primary interface eth0 works perfectly, and is in fact how I am ssh'd in. However, the secondary interface eth1 is not transmitting. I can see this in ifconfig output, for example, where the TX field is always 0. However, it is receiving, and tcpdump shows ARP requests coming from the ISP's gateway on the other side. The interface is physically connected to a Siemens BSTU4 modem, configured by the ISP. The link is properly set to 10MBps and full duplex, without negotation, as the ISP requested. A small /30 subnet is configured. For the sake of anonimity, let's say the machine is 3.3.3.2/30, and the ISP's gateway .1. The machine has no firewall settings whatsoever. Even running something like arping -I eth1 3.3.3.1, and running tcpdump alongside, shows no traffic whatsoever being transmitted on the interface. (But the other side keeps steadily sending ARP requests, and that is all that can be seen.) What could be causing this? Here's some output, anonymized, which may hopefully help: $ ethtool eth1 Settings for eth1: Supported ports: [ TP ] Supported link modes: 10baseT/Half 10baseT/Full 100baseT/Half 100baseT/Full 1000baseT/Full Supports auto-negotiation: Yes Advertised link modes: Not reported Advertised auto-negotiation: No Speed: 10Mb/s Duplex: Full Port: Twisted Pair PHYAD: 1 Transceiver: internal Auto-negotiation: off Supports Wake-on: d Wake-on: d Link detected: yes $ ip link show eth1 3: eth1: <BROADCAST,MULTICAST,UP> mtu 1500 qdisc pfifo_fast qlen 1000 link/ether 00:15:c5:xx:xx:xx brd ff:ff:ff:ff:ff:ff $ ip -4 addr show eth1 3: eth1: <BROADCAST,MULTICAST,UP> mtu 1500 qdisc pfifo_fast qlen 1000 inet 3.3.3.2/30 brd 3.3.3.3 scope global eth1 $ ip -4 route show match 3.3.3.0/30 3.3.3.0/30 dev eth1 proto kernel scope link src 3.3.3.2 default via 10.0.0.5 dev eth0

    Read the article

  • Grizzly server - works with IP, but not with domain name

    - by Hitchhiker
    I'm hosting a grizzly web service on a Windows 7 Pro machine (embedded in a regular Java process), and it is binding to http://my-domain-name. When trying to hit the service from another machine, requests to http://my-domain-name fail (fiddler shows error code 502), but requests to http://my-ip work. When the service runs on a Windows Server 2008 machine, this doesn't happen (both requests succeed). What could be the issue?

    Read the article

  • Well....a ghost lives in my server...

    - by tsgiannis
    Hello to every body and greetings from Greece I have a rather unusual situation and i am running out of ideas. I have this old server (IBM x205 - P4 2.4Ghz,3xSCSI 36gb) and i was about a year ago i decided to use it as an additional domain controller,fax & file server...for this task i had a Delock 70154 SATA card along with 2x320 SATA II HDDs Everything was going super smoothly until about 3 weeks ago. I was on a trip and i was informermed when i got back that the server was found frozen...well i considered it was a glitch since a simple powerdown power up fixed everything.Again 2 weeks ago another freeze situation....it got suspicious but again power down power up everything was running.... Last again it frozen and when i power it up it came with a message that the Domain services could not start due to NTDS corruption....booting in safe mode revealed that there was an issue with the SATA Raid (degraded). After a lot of searching i degraded the server...cleaned Active Directory and pulled both HDDs out (one of these was really BAD ) and recovered my files (i had some problems with ho Delock handles the redudant HDD). Right now my server is vanilla simple...with only what the factory installed and here is where the fun begins. Everyday when i arrive at the office i find this particular machine dead..and i mean total dead...just a black screen and nothing else...the cpu fan is working ,the psu is working .keyboard and mouse are dead(they also lock my kvm) ...network is dead.... the machine is DEAD. I power it down forcilly ...i power it up and for the 8 hours i am in the office it works,either idling or running some kind of diagnostic...when i leave the office after some time..it maybe half an hour ...it maybe 4 hours the machine dies...at least this is the information the event log shows (" the previous shutdown at xx:xx:xx was unexpected) Well i must admit i am runnig out of ideas.... I have tried Memtest....nothing Passmark burn in test.....nothing Carefull study of the event log.....nothing Set Instead of restart..BSOD....nothing Power sceheme to sleep...all set to never. i know there are a lot of other tools that heavily stress a machine like occt but .... the machine is old...today i will give them a try nevertheless ..... One idea is to reformat it...but ... i really like to find what is causing this because i could get to a situation that everything is working for a while and kaboom...one day again is dying.. I really need a helping hand and every opinion / idea is well welcomed.... iknow the obvious solution is to never leave the office but....i have a life...sory server...:) P.S this situation with the machine dying some time after is going on for about one week...everyday i would set either the RAID to rebuild....or to copy/recover files and while everythig was working

    Read the article

  • Centos swap cache memory leak

    - by user30008
    We have a image server that keeps running out of memory and crashing, we thought there was a hardware issue with the machine because the code base has not changed and this is a new issue. We brought a new machine online with newer kernel and fresh centos 5.4 install and just brought online one subdomain and the exact same error is occurring on the new machine. How should I try and troubleshoot this issue.

    Read the article

  • VMWare Server with multiple processors

    - by user43046
    I have a new linux machine with two Core Duo CPUs. However, VMWare Server only recognizes one. In the host summary it shows: Processors: Intel(R) Core(TM) i5 CPU 750 @ 2.67GHz 1 CPU x 2 Cores On another machine, it shows: Intel(R) Core(TM)2 Quad CPU Q6700 @ 2.66GHz 1 CPU x 4 Cores This machine also has two CPUs. Howcome VMWare is not seeing all the CPUs?

    Read the article

  • TechEd 2014 Day 3

    - by John Paul Cook
    There is some confusion about durability of data stored in SQL Server in-memory tables, so some review of the concepts is appropriate. The in-memory option is enabled at the database level. Enabling it at the database level only gives you the option to specify the in-memory feature on a table by table basis. No existing tables or new tables will by default become in-memory tables when you enable the feature at the database level. If you choose to make a table an in-memory table, by default it is...(read more)

    Read the article

  • Stairway to T-SQL DML Level 10: Changing Data with the UPDATE Statement

    Unless you are working on a reporting-only application you will probably need to update tables in your SQL Server database. To update rows in a table you use the UPDATE statement. In this level we will be discussing how to find and update records in your database, and discuss the pitfalls you might run into when using the UPDATE statement. Is your SQL Database under Version Control?SSMS plug-in SQL Source Control connects SVN, TFS, Git, Hg and all others to SQL Server. Learn more.

    Read the article

  • Another Update to SQL Server Configuration Scripting Utility

    - by Bill Graziano
    I’ve been gradually adding features to my utility that scripts the configuration of a SQL Server.  Since my last post I’ve added the following features: Skip any encrypted object in a database Script alerts, alert notifications and operators Script audits Always script model, master and msdb to capture any user-defined objects in those databases Logins are now scripted so that everything for a login is grouped together. There’s a second section in the logins that handles default databases.  In many cases a login’s default database is a mirror target and can’t be set.  This is now handled gracefully.  It also includes a separate section for all default databases so those can be quickly set in the event of a disaster. Script credentials Script proxy accounts Script database mail My goal is still to get everything outside a database scripted.  This release is enough that I can keep my mirror target servers in sync with their principals.

    Read the article

  • Distributed website server redundancy

    - by Keith Lion
    Assume a website infrastructure is very complicated and is fully distributed (probably like most large web companies). Am I right in thinking that although there are all these extra web servers to handle multiple client requests, there is still a single "machine" whereby users must enter? I am guessing this machine will be the one physically associated to the IP address? I ask because I need to know whether, in places where distributed systems exist, there is still a single point of failure- usually the control node or, in this example, the machine connected to the public internet? Surely there cannot be two machines connected to the internet, as they would have to have different IP addresses? This "machine" may not be a server per se, but maybe it is a piece of cisco equipment. I just need to know whether, in the real world, these distributed systems still have a particular section where they depend on the integrity of one electronic device?

    Read the article

  • Mirroring: what happens if principal loses contact with both mirror and wittness?

    - by TiborKaraszi
    Imagine a database mirroring setup where you have two LANs with a WAN link in between. Let's call them site A and site B. Say that principal is currently in site A, and both mirror and witness are in site B. I.e., we are running database mirroring with a witness, and assuming we are running safety FULL (synchronous), we have auto-fail over. Now, what is really fail over when it comes to mirroring? the simple answer is that the mirror will finish the recovery process (UNDO) and make the database available....(read more)

    Read the article

  • How to reach a Global Scope IPv6 host?

    - by Vaibhav Bajpai
    I have setup DNS64+NAT64 on a machine with 2 interfaces: eth0: public IPv4 address (connected to outside world) eth1: global scope IPv6 address: 2001::/64 I can successfully use ping6 google.com on this machine. Now I want to connect my MacBook to this machine by making it an IPv6-only client and perform some tests, but I do not have an IPv6 address assigned on this MacBook. I'm wondering, how should I manually assign one so as to route all my IPv6 traffic (I will disable IPv4 on my MacBook) to this machine, which will be picked up by DNS64+NAT64 to be converted to IPv4 requests and sent to the outside world?

    Read the article

  • Working with Git on multiple machines

    - by Tesserex
    This may sound a bit strange, but I'm wondering about a good way to work in Git from multiple machines networked together in some way. It looks to me like I have two options, and I can see benefits on both sides: Use git itself for sharing, each machine has its own repo and you have to fetch between them. You can work on either machine even if the other is offline. This by itself is pretty big I think. Use one repo that is shared over the network between machines. No need to do git pulls every time you switch machines, since your code is always up to date. Never worry that you forgot to push code from your other non-hosting machine, which is now out of reach, since you were working off a fileshare on this machine. My intuition says that everyone generally goes with the first option. But the downside I see is that you might not always be able to access code from your other machines, and I certainly don't want to push all my WIP branches to github at the end of every day. I also don't want to have to leave my computers on all the time so I can fetch from them directly. Lastly a minor point is that all the git commands to keep multiple branches up to date can get tedious. Is there a third handle on this situation? Maybe some third party tools are available that help make this process easier? If you deal with this situation regularly, what do you suggest?

    Read the article

  • How much resources are used during a RDP connection?

    - by KronoS
    Having recently installed a RDP app on my mobile device and got me wondering: this small device which has a limited amount of resources compared to the PC I'm connecting to, is still able to operate my remote machine to almost full capacity. With that, how much of the local machines resources are actually used in the connection to the remote machine? Also how much of the resources are used on the remote machine?

    Read the article

  • Get script for every action in SQL Server Management Studio

    I am always conscious to keep a record of all operations performed on my database servers. Operations through T-SQL in an SSMS query pane can easily be saved in query files. For table modifications through SSMS designer I have predefined setting to generate T-SQL scripts. However there are numerous database and server level tasks that I use the SSMS GUI and I would like to have a script of these changes for later reference. Examples of such actions through the SSMS GUI are backup/restore, changing compatibility level of a database, manipulating permissions, dealing with database or log files or creating/manipulating any login/user. I am looking for any way to generate T-SQL code for such actions, so that it may be kept for later reference

    Read the article

  • Unable to view 2 local sites over network

    - by gentrobot
    I have 2 websites running on my local machine that I'd like to view from other machines on the same network. For /etc/apache2/sites-available/site1.com: <VirtualHost *:80> ServerName site1.com DocumentRoot /var/www/answers/app/webroot DirectoryIndex index.php <Directory "/var/www/answers/app/webroot"> Options FollowSymLinks AllowOverride All Order allow,deny Allow from all </Directory> </VirtualHost> For /etc/apache2/sites-available/site1.com: <VirtualHost *:80> ServerName site2.com DocumentRoot /var/www/answers2/app/webroot DirectoryIndex index.php <Directory "/var/www/answers2/app/webroot"> Options FollowSymLinks AllowOverride All Order allow,deny Allow from all </Directory> </VirtualHost> I have added 2 entries in the /etc/hosts file as: 127.0.0.1 site1.com 127.0.0.1 site2.com Now, when I point the browser on my machine to site1.com, it shows me the first site and pointing the browser to site2.com, it shows me the second site. However,when I type in the local IP of my machine in the browser, it always shows site2. How can I change it to switch between site1 and site2 ? Is there a way that I can view both the sites form another machine (esp. mobile devices over wireless network) ?

    Read the article

  • Remote connection to Mysql not working

    - by Fillipe Silva
    We have an application running with CodeIgniter and MySQL in cestacerta.com. Ther are two versions: One that is in production and it works perfectly, and one that is our Development version. This runs locally on my machine and the other developers. The development version needs to remotely connect to the database, and it rather suddenly stopped working. I've given permission to access any IP settings on the server. I can access the database through the MySQL Gui Tools. I have tested access to several different codes, including a newly downloaded version of CodeIgniter and always got the same error: "A Database Error Occurred Unable to connect to your database server using the provided settings. Filename: C: \ xampp \ htdocs \ cestacerta \ system \ database \ DB_driver.php Line Number: 119 " What troubleshooting steps can I take determine if the error is on our workstations (which all have the same error) or on our server.

    Read the article

  • Full/Differential backup - what's used to determine the differential backup content?

    - by gernblandston
    Let's say I have a 'MyDB' SQL Server 2005 database (simple recovery) in which I do a Full backup on Sunday, and Differentials every other night BACKUP DATABASE [MyDB] TO DISK = N'c:\Database Backups\MyDB\MyDB_Full.bak' WITH NOFORMAT, INIT, NAME = N'MyDB.BAK', SKIP, NOREWIND, NOUNLOAD, STATS = 10 and BACKUP DATABASE [MyDB] TO DISK = N'c:\Database Backups\MyDB\MyDB_Diff.bak' WITH NOINIT, DIFFERENTIAL, NAME= 'MyDB.BAK', STATS= 10 What does the differential backup process use to decide what data gets backed up on the differential nights? Does it need the mydb_full.bak file to do its business? If I wanted to save disk space, could I zip up the mydb_full.bak file to a .zip file after it's created without adversely affecting the differential backups, and if I needed to restore, just unzip the full backup before starting?

    Read the article

  • Problems with Startup after Script runs to Shutdown Windows 7 PC

    - by Toby Allen
    I have a simple .bat script to shutdown a kiosk style machine in our work shutdown -h -f this hibernates the machine and I have it running each evening as a scheduled job. However my problem comes when I turn the machine on again. The first time I press power it starts up and then immediately hibernates again, like it runs the script a second time. the second time I press power it stays up. Any suggestions on how I can avoid this?

    Read the article

  • Can't boot from CD

    - by Jason Swett
    I installed Ubuntu on a PowerBook G4 and it only works about 10% of the time. For this reason I decided to re-install OS X, but my machine won't boot from CD. It also won't boot Ubuntu the vast majority of the time, so I'm kind of screwed. After a ton of tries, I finally got my machine to boot a Linux command line. Is there anything I can do at this point to make my machine more runnable, just for the sake of installing OS X?

    Read the article

  • Can't boot from CD

    - by Jason Swett
    I installed Ubuntu on a PowerBook G4 and it only works about 10% of the time. For this reason I decided to re-install OS X, but my machine won't boot from CD. It also won't boot Ubuntu the vast majority of the time, so I'm kind of screwed. After a ton of tries, I finally got my machine to boot a Linux command line. Is there anything I can do at this point to make my machine more runnable, just for the sake of installing OS X?

    Read the article

  • Azure website that talks to third party services

    - by Andy Frank
    I have website that crawls data from many third party services when user browse to webpage. This can be really slow because I hit third party server and process returned data before showing it to user. I am hosting website on Azure (shared mode). I am thinking to improve my implementation. Here is what I am thinking... Run a service that crawls data from third party services, process it and then store it in database. when user browse to my site, my site pulls data from database and display them to user. But above solution is not clear to me. Should I have normal service or wcf service? If wcf service then should website talk to database or wcf service (that can access data from database)? If normal service then how can I deploy on Azure?

    Read the article

  • How can I have sound output before logging in?

    - by ??O?????
    I have a machine (Ubuntu 11.10) that I would like to have it play audio (typically through an amplifier), but the machine should be headless (where its final placement will be). I will control what is played through ssh. However, there is no sound output until I login to the graphical console. At first I thought it was an issue with pulseaudio, so I promptly removed it to use the default ALSA, but I have the same issues. I ssh to the machine, I run alsamixer and get the typical cannot open mixer: No such file or directory error (while /proc/asound/cards display correctly what I have). If I login on the graphical console, alsamixer works fine in the ssh session, and I have sound output. I logout, and then alsamixer stops working. So something runs (Xsession perhaps?) when I login that enables sound output, and gets disabled when I logout. I remember in older versions of Ubuntu, there was a drum roll when the machine showed the login screen; that is not the case anymore. Perhaps if I somehow can enable that drum roll, I'll have fixed my problem too. In any case, the question I ask is what the title says.

    Read the article

  • How to wipe the slate clean on an Oracle installation?

    - by nw
    After a basic, successful installation of Oracle 11g, I ran dbca again to enable Enterprise Manager on my database. The operation hung at around 67% for half an hour or more, so I clicked Stop to abort the operation. Things seemed to wrap up cleanly, EM was working, all was well with the world. Then I started getting this dreaded error upon any attempt to connect in SQL*Plus: ORA-12154: TNS:could not resolve the connect identifier specified I thought perhaps the database had been corrupted due to the earlier aborted operation, so I ran dbca again and deleted the database. Then I attempted to create a new database in its stead, using a clone of the template created the first time around. Unfortunately, the clone database operation fails at 50% with the exact same error: ORA-12154: TNS:could not resolve the connect identifier specified How can I clean up the mess I have created, short of reinstalling Oracle entirely from scratch?

    Read the article

< Previous Page | 491 492 493 494 495 496 497 498 499 500 501 502  | Next Page >