Search Results

Search found 19586 results on 784 pages for 'machine instruction'.

Page 239/784 | < Previous Page | 235 236 237 238 239 240 241 242 243 244 245 246  | Next Page >

  • Remote Desktop Multi Monitor Connection

    - by user196039
    This question may be a bit unusual but I think it's an interesting angle: Why does a remote desktop connection to my server with all four of my 30 inch (@2560x1600) monitors work, even though the server doesn't have a graphics card installed? I guess the graphics are perhaps not really rendered on the server? What exactly happens there so that this works? Searching for this I mostly found "how to enable multi monitor support" but not an answer to my question yet. Any links/documentation I can read to understand this better? (To get all four monitors working well on my local machine the one graphics card I had wasn't even enough, so I have two graphics cards on the local machine now) The operating systems are Windows 7 home premium and it's a Windows Server 2008 R2.

    Read the article

  • How do I edit files in the console when connecting to windows 7 via ssh?

    - by Alex Waters
    I am using tunnelier client and server to connect to a windows machine. I can get in and have access to all of the files on the computer. I have vim installed on that windows machine, but I can't seem to edit anything via the DOS command line. I also tried editing in notepad, but nothing happens when I enter the command. I think this might be the part where DOS doesn't behave like bash. Would I need to setup cygwin / openssh to accomplish this? (boo, tunnelier is so easy) Thanks! p.s. I know I could just use sftp and edit files that way, but it feels dirty.

    Read the article

  • OSX : Setup for filestorage in medium business

    - by Franatique
    In our office every machine runs OSX. In search of an ideal storage and sharing solution we decided to let OSX Server handle all account information and auth requests whereas an 7TB QNAP provides NFS shares. All shares are published as mounts in the companywide LDAP. As it turns out, handling permissions in this situation is very clumsy (e.g. inherit permissions on newly created files). Unfortunately using NFS4 in combination with ACLs did not solve the problem. As a possible solution I set up a iSCSI connection between QNAP and the machine running OSX Server which in turn serves the LUN as AFP share. Permission handling works like a charm for this setup. Although I am a bit concerned about the performance of this setup. As we are a fast growing company we expect the solution to serve at least 100 clients while using files aprox. above 100MB each. Are there any known drawbacks of this solution?

    Read the article

  • Is it safe to sync a Firefox instance in both directions?

    - by java.is.for.desktop
    Hello, everyone! I want to use some tool (not decided yet, which one) to sync a Firefox instance (more exact: user directory) between two machines. (EDIT: I want really to sync everything in the user-directory) Would you assume, that this is safe? What would happen, if some files are newer on the one machine, and some other files are newer on the other machine? Could this lead to inconsistencies, when, let's say, there are some inter-file references? If general, I don't have good experience with syncing application in both directions. Most applications seem not to be suited for this.

    Read the article

  • Why does it sometimes take over a minute for Apache2.2 to serve a static page?

    - by Jason Lamoreux
    We have a Windows Server 2003 machine running Apache2.2. Most of the time there is no load on the server, but we have a notification program on 3400 PC's that can request a small web page that plays a 64KB .wav file. When an event occurs those 3400 PC's all request the web page over the course of 3 minutes. On a few machine we saw the browser sit in the "connecting" state for a little over a minute before the page painted. What is happening, and how can we speed this up?

    Read the article

  • Too many files open issue (in CentOS)

    - by Ram
    Recently I ran into this issue in one of our production machines. The actual issue from PHP looked like this: fopen(dberror_20110308.txt): failed to open stream: Too many open files I am running LAMP stack along with memcache in this machine. I also run a couple of Java applications in this machine. While I did increase the limit on the number of files that can be opened to 10000 (from 1024), I would really like to know if there is an easy way to track this (# of files open at any moment) as a metric. I know lsof is a command which will list the file descriptors opened by processes. Wondering if there is any other better (in terms of report) way of tracking this using say, nagios.

    Read the article

  • ssl problems between two tomcat servers

    - by user1058410
    I've got one tomcat 7 server(kind of like a web server) trying to talk to a tomcat 6 server(acting as a document server on another machine) over ssl and keep getting this error java.security.cert.CertificateException: No name matching rarity64 found. Where rarity64 is the name of the document server. I've tried exporting keys from both tomcats keystores and importing them into the others keystores using java keytool. I've even tried adding them to the other machines cacerts keystore. I've also used internet explorer to import both keys into the other machine. But nothing I try works. If it matters the real webserver is IIS 7.5, which the tomcat "webserver" talks too with arr, and they don't use SSL. But the problem seems to between the two tomcat servers

    Read the article

  • Hyper-V Manager - Host Access During a Catastrophe

    - by LonnieBest
    How can I ensure that I can always have Hyper-V Manager access to a Hyper-V server, even in the event that the Active Directory Server is down (in a domain-login environment)? Background: The one that came before me, set up the company's servers as virtual machines on top of a host running Hyper-V Server 6.1 (7601) Service Pack 1. For managing Hyper-V, he installed Window 7 onto a virtual machine (run on the same host) with Hyper-V Manager installed. When the (virtual) Active Directory server (run on this same host) is rebooted, during that reboot, I'm unable to RDP into the Windows 7 virtual machine, and I'm therefore unable to access Hyper-V Manager when the Active Directory server is down. I suspect I can't login because I can't authenticate with the Active Directory Server. I'm going to install Hyper-V Manger onto some addition manager's workstations, but how can I ensure they'll have access in a catastrophe where Active Directory authentication isn't possible?

    Read the article

  • Windows 8 Task Manager RAM Usage Accuracy

    - by user264892
    The new Task Manager has a great UI in windows 8, however, there are some discrepancies in the data I can not account for: Machine: 8 GB of total ram. (This is a physical machine, not a virtual) The processes tab shows 45% of Memory utilized. The listed process do not add up to 3.5 GB of RAM, but instead add up to 0.948 GB. There is no "processes for all users" option. The performance Tab Shows: In use : 3.6 GB Available: 4.4 GB Committed : 4.1 /9.2 GB Cached: 3.7 GB Paged Pool: 376 MB Non-paged pool: 135 MB My reading of this says I have ALOT of "cloaked" processes running some where eating my ram. How do I interpret this data and how do I verify it?

    Read the article

  • How do I configure upstart to run a script on shutdown when the process takes longer than 10secs?

    - by Tiris
    I am running ubuntu 11.10 in a virtual machine (VirtualBox) to learn more about development in linux. I am using a git repository to save my work and have written a script to bundle my work up and save it to the shared folder for use while the virtual machine is not running. I would like to automatically run this script before shutdown so that my work is always available if the vm is off (currently I have to manually run the script). I don't know if upstart is the best way to accomplish this, but this is the config that I wrote as a test: description "test script to run at shutdown" start on runlevel [056] task script touch /media/sf_LinuxEducation/start sleep 15 touch /media/sf_LinuxEducation/start-long end script pre-start script touch /media/sf_LinuxEducation/pre-start sleep 15 touch /media/sf_LinuxEducation/pre-start-long end script post-start script touch /media/sf_LinuxEducation/post-start sleep 15 touch /media/sf_LinuxEducation/post-start-long end script pre-stop script touch /media/sf_LinuxEducation/pre-stop sleep 15 touch /media/sf_LinuxEducation/pre-stop-long end script post-stop script touch /media/sf_LinuxEducation/post-stop sleep 15 touch /media/sf_LinuxEducation/post-stop-long end script The result is that only one touch is accomplished (the first touch in pre-start). What do I need to change to see one of the touches after the sleep to work? Or Is there an easier way to get this accomplished? Thanks in advance.

    Read the article

  • External disk not working with Mac OS X 10.5

    - by ctx2002
    Hi: My OS is 10.5.7. I have an external hard disk, formatted as Mac OS Extended (Journaled), connected to my Mac machine through USB cable. I have used it for a while, but now it is not working. Error showed in /var/log/system.log. USBF: 2521.384 AppleUSBEHCI[0xa515800]::Found a transaction which hasn't moved in 5 seconds on bus 0xfd, timing out! (Addr: 0, EP: 0) ..... ...... kernel[0]: USBF: 2588.616 [0xa6baa00] The IOUSBFamily was not able to enumerate a device. Luckily, in my work place there is a Mac with OS 10.4, so I had plugged the disk into the 10.4 machine and it worked!!! I did checked "System Profiler" on both machines (10.5, 10.4). Under usb section, for 10.5 it shows USBEHCI, under 10.4, it shows USBOHCI. I am not a expert on USB Device , can any one tell me why 10.5 cause the problem. it was worked under 10.5

    Read the article

  • Hyper-V R2: Need help with disk structure

    - by MojoDK
    Hi all, I'm going to use the free (non gui) version of Hyper-V R2. In my new server I have 8 disks in total (for Hyper-V R2 installation and virtual machine). Atm I'm going to run a single virtual machine, with following tasks: Windows Server 2008 R2 x64 File/Print SQL Server My question is ... with my 8 disks in the server, which disks should contain wich data? Should I install "Hyper-V R2" and VM's drive c on same physical disks? Should I use raid 1 or 5? With the above tasks, how would you structure the disks? Hope you know what I mean (I'm not english, so it's difficult to explain). Thanks!!! Mojo

    Read the article

  • Ubuntu 11.10 Random reboots how to find root cause?

    - by Indigo42
    All, This issue started with 11.04, I upgraded to 11.10 this week in hopes it would correct it, but after a few days it happened again.. A few times a day, randomly, my monitors will go black and the machine will come back in the post screen and boot up again. Sometimes it boots up properly, sometimes it boots to disk scan (I understand why...the machine just died with a bunch of stuff open) and sometimes it takes several cycles of this to come up. I connected to the system with my laptop and tailed dmesg, syslog, kern.log, and Xorg.0.log. I also ran a cat /proc/kmesg. Nothing is showing up in any of these log files when it happens..it's just like someone has pulled the plug. I just got through running 4 hours of the memory test with no errors. I have a AMD Phenom 9850 Quad-Core Processor and Nvida 9500GT graphics card. This system has been stable for like 4 years. I'm thinking it might be a hardware problem, but how do I find out? Thanks,

    Read the article

  • Writing directory: permission denied even though dir seems to be chmodded correctly

    - by Aron Rotteveel
    I am having some trouble creating files in directory on my Ubuntu machine: I added myself to the www-data group in order for me to easily edit stuff in my /var/www dir on my development machine. stat /var/www shows the following: File: ‘/var/www’ Size: 4096 Blocks: 8 IO Block: 4096 map Device: 808h/2056d Inode: 142853 Links: 3 Access: (0775/drwxrwxr-x) Uid: ( 33/www-data) Gid: ( 33/www-data) Access: 2010-12-30 16:03:18.563998000 +0100 Modify: 2010-12-30 16:02:52.663998000 +0100 Change: 2010-12-30 16:03:13.111998001 +0100 Still, it is impossible for me to create anything below /var/www (the only way for it to work is to chmod it to 777. What am I missing?

    Read the article

  • Upgraded from ubuntu 12.04 to 12.10 issues

    - by ubuntu novice user
    I recently upgraded 12.04 to 12.10 ubuntu and then all hell broke loose. Being more specific, I have a Compaq machine and its hard disk is partitioned into 3 parts, so when I installed Ubuntu 10.04, I installed it in windows, since then have upgraded with each new ubuntu release via the update manager without any problems. I have installed the 64 bit versions. 12.10 downloaded via update manager, and initial downloading of packages was without problems, however, as it tried to install the packages, error messages appeared. The first was one about missing lib files, but I clicked to continue, since I am a relative novice on ubuntu. It continued, and during the restart process, when it was powering down, the computer hanged on the shutting down bit without rebooting for more than half an hour, so I manually shut down the machine and restarted it. Then a new error message appeared stating could not find disk, and I hit the manual fix option, and it now boots to an empty ubuntu desktop with my wallpaper but no launcher and the graphics appears as if this was put on a 640 x 480 resolution, and the screen no longer fits onto my 19" LCD. I had to use Ctrl-Alt-T to log out and then restart from there. How can I resolve this issue. Please help!

    Read the article

  • What is your list of programs to install to Windows after a fresh install?

    - by marked
    I am looking for others' list of program that absolutely must be installed to a fresh install of Windows before going any further. I hope to compile a list here to use as reference for all new Windows installs/restores. Automating this list of programs installation is the next step. Some great lists so far! I will continue to monitor this and then make a list of everything I use. Another thought, has anyone come up with a way to automate this? Possibly going to each programs download page, downloading the most recent version, and installing? I may way on this as well, any ideas would be appreciated! As far as the type of machine, most of mine are hybrids i.e- server, workstation, development machine. So all of the above!

    Read the article

  • mongodb eating 48G in 1min

    - by ledy
    In mongodb i work with this collection: Size 55.93g Data Size 39.82g Storage Size 41.08g Extents 53 Indexes 4 Index Size 9.64g It takes few seconds of mongdb being up with this single collection and all 48GB RAM on the dedicated server are gone. That's worse because there is also a mysqld+nginx/fcgi on this machine which should be allowed to use at least 24GB together. I.e. remaining 24GB, enough for the mongod! However, it does not share in a fair way. Everybody says that the memory for mongod is managed by OS and releases unneccessary space for other processes if they demand RAM. On my machine it is not releasing RAM. What's wrong? free total used free shared buffers cached` Mem: 49559136 49403908 155228 0 57284 47247564 -/+ buffers/cache: 2099060 47460076 Swap: 8008392 164 8008228

    Read the article

  • 3.2.0-31-generic version doesn't boot

    - by user92526
    12.04 used to work just fine a month ago. Since August, I installed all of the available updates. Since then my mac-pro is acting weird. If I restart the computer, the screen gets stuck in the purple back ground without any texts. Then if I restart again, I can choose what versions of linux I want to run. If I select 3.2.0-31-generic, the machine gets stuck in a state of blinking cursor. If I boot again to run 3.2.0-31-generic recovery mode, the machine gets stuck in a "..... memory freed End of Stack" mode. I can't boot through this kernel. I have to choose an older version of linux like 3.2.0-12-generic to boot into my mac pro, but I have too boot twice to get into this version. I was wondering --if you could provide me a solution to boot the 3.2.0-31-generic version. -- if you could provide me a solution to boot into any version the first time I restart. (Usually I have to kill the first restart) --- If there is an option in linux to choose what version I want to run every time I restart so that I don't have to choose everytime. Thank you

    Read the article

  • Can't delete profile

    - by generic_noob
    Hi All, I have a client machine (XP SP3) which used to be connected to a windows 2003 domain, but the server has since gone down due to a hard drive failure, however the profiles still remain in 'documents and settings'. I have a local admin account on the same machine (in the administrator's group), except when i try to remove the profiles manually from System Properties - Advanced - User Profiles - Settings, the delete button is disabled. Also, Windows prevents me from deleting or renaming the user's profile folder as well, due to a lock with ntuser.dat Any thoughts on this would be greatly appreciated, Cheers!

    Read the article

  • Ubuntu: Resize the root LVM(2?) partition

    - by user12259
    I have an Ubuntu virtual machine running in VirtualBox 2.2.4, and I created it on an 8gb virtual disk which is too small. So, I am trying to increase the size of the disk. So far, I have done this: Created a new larger virtual disk Added the 2nd disk to the machine Used CloneZilla to clone the first disk onto the 2nd disk Removed the first disk Booted up off the 2nd (larger disk) But now I'm still stuck with an 8gb partition on my new 100gb virtual disk. Whats the easiest path from here to having a 100gb partition? :) I gather GPart can resize partitions, but it doesn't seem to support LVM2 partitions, which mine seems to be. thx Alex

    Read the article

  • tab completion for service command on debian

    - by markus
    I have two systems with debian squeeze installed. On one system when I type: service <TAB> it shows me all available service (from /etc/init.d) on the other system it shows me all files from the current directory. Does anyone know which setting changes that behaviour ? UPDATE: The file /etc/bash_completion.d/service was missing. I copied it from the machine where it is working. If I type complete -p | grep service it shows me: complete -F _service service On the machine where it is not working that command shows me nothing. I executed complete -F _service service in the command line, after that, the command service <TAB> shows me: service -su: completion: function `_service' not found this function is defined in the service file I recently copied, for some reasons it can't be found ...

    Read the article

  • vSphere 4 - how can I cancel a file copy in progress?

    - by DrStalker
    VMware vSphere 4 SAN storage with multiple data-stores No vCenter I shut-down a virtual machine and using the data-store browser did a copy/paste to copy the VM to a new datastore with additional space. The file copy performance was very poor, and due to time constraints I decided to cancel the copy task. However the copy task showing in the vsphere client can not be cancelled; the cancel option is disabled. Currently I am not able to start the machine in it's original location as the disk files are locked for the copy. How can I abort the copy? I tried deleting the target directory but this did not abort the copy task.

    Read the article

  • Should Production Windows Web Servers (IIS & SQL) be in a domain?

    - by tlianza
    We have a few web servers and a few database servers. To date, they've been standalone machines that are not part of a domain. The web servers don't talk to each other, and the web servers talk to the database servers via SQL Auth. My concern with putting the machines in a domain together were added complexity - it's one more "thing" running, and doing "things" that could go wrong. risk - if a domain controller fails, am I now putting other machines at risk? However, in certain scenarios it does seem convenient for them to be on a domain, sharing credentials. For example, if I want to give the "services" control on one machine access to another machine (because Remote Desktop craps out) I need to go in and assign privileges on multiple machines - something that I believe Active Directory and Domain Accounts set to simplify. My question: I'm sure there are things I'm not considering here. Is there a best practice?

    Read the article

  • How to exchange the HDD of a MacBook Pro?

    - by Another Registered User
    I've bought an Solid State Drive (SSD) for my MacBook Pro, and now I need to exchange it somehow. Would this strategy work? 1) Create an backup with Time Machine (Snow Leopard) 2) Then replace the old HDD 3) Insert the new HDD 4) Install Snow Leopard (same version as previously used) 5) Open up Time Machine, and recover from the last backup I'm not sure about how to do the last part. Is that hard? What are the neccessary steps? Or is there a better way? Maybe I don't need to re-install Snow Leopard completely? Maybe the Install CD already offers an option to recover from Backup?

    Read the article

< Previous Page | 235 236 237 238 239 240 241 242 243 244 245 246  | Next Page >