Search Results

Search found 67158 results on 2687 pages for 'work life'.

Page 506/2687 | < Previous Page | 502 503 504 505 506 507 508 509 510 511 512 513  | Next Page >

  • Should I choose KVM/XEN over OpenVZ or use them together?

    - by Krystian
    I've got a dual xeon e5504 server, with [for now] only 8GB of ram. Storage is'n impressive either: 3x 146GB sas in raid5 + 500GB sata drives. Currently it works as a development server, but it's over speced for our needs and since our development methods changed through last 2 years we decided it will work as a production system for some of our applications + we would like to have a separate system for testing/research. Our apps are mainly web apps deployed on tomcats [plural as some of the apps require older versions] and connected to Postgres. I would like to have a production system, where only httpd+tomcat+db are setup and nothing else runs there. Sterile system. Apart from that, I would like a test system, where I can play with different JVM settings, deploy my test apps, play with tomcat/httpd settings and restart them without interfering with the production system. Apart from that, I would like to be able to play with different linux flavors, with newer kernels to test how they work etc. I know, this is not possible with OpenVZ and I would have to choose KVM for that. I am thinking about merging the two, and setting up a KVM to be able to work with different systems [linux only to be frank] + use openVZ to setup separate machines for my development needs. I would simply go with that, but reading here and there about the performance impact full virtualization has over containers and looking at the specs of my server makes me think twice about it. I don't want to loose too much performance, especially because of the nature of my apps [few JVMs running at the same time]. It will be my first time with virtualization, apart from using desktop virtualbox/vmserver. Although I am a fast learner I don't want to mess with the main system so much that it will break the production apps or make them crawl. Although they are more or less internal apps and they don't produce much load, they need to be stable. I've read, that KVM host is a normal linux installation and it allows to run normal processes on it. If that is so, does it allow to run openVZ as well? I mean... can I have KVM and OpenVZ running on the same system/kernel? Or do I have to setup another system to run OpenVZ containers? How much performance impact can this have for me? Will my hardware suffice? oh and one more thing... unfortunately I'm quite limited with the funds... I'm looking for a free solution only :/

    Read the article

  • Pushing Large Files to 500+ Computers [closed]

    - by WMIF
    I work with a team to manage 500-600 rented Windows 7 computers for an annual conference. We have a large amount of data that needs to be synced to these computers, up to 1 TiB. The computers are divided into rooms and connected through unmanaged gigabit switches. We prepare these computers ahead of time with the Windows installation and configuration, plus any files that we have available to us before we send the base image in for replication by the rental company. Every year, we have presenters approach on site with up to gigs of data that need to be pushed to the room that they will be presenting in. Sometimes they only have a few files that are small sizes, such as a slide PDF, but can sometimes be much larger 5 GiB. Our current strategy for pushing these files is using batch scripts and RoboCopy. For the large pushes, we actually use a BitTorrent client to generate a torrent file, and then we use the batch-RoboCopy to push the torrent into a folder on the remote machines that is being monitored by an installed BT client. Often times, this data needs to be pushed immediately with a small time window. We have several machines in a control room that are identical to the machines on the floor that we use for these pushes. We occasionally have a need to execute a program on the remote machines, and we currently use batch and PSexec to handle this task. We would love to be able to respond to these last minute pushes with "sorry, your own fault", but it won't happen. The BT method has allowed us to have a much faster response time, but the whole batch process can get messy when there are multiple jobs being pushed. We use Enterprise Ghost for other processes, and it doesn't work well in this large of scale, plus it is really quite expensive for a once-a-year task like this. EDIT: There is a hard requirement that the remote machines on the floor are running Windows. The control machines do not have a hard OS requirement. I would really like to stay away from Multicast because of complications with upstream routers. Is Multicast or BitTorrent the better way to go on this? Is there another protocol that might work better?

    Read the article

  • How can I change shell in Ubuntu?

    - by Sanoj
    I have a Ubuntu Server VPS and I want to use /bin/bash/ as my shell. How can I change my shell? I have root access but I don't work as root. So dash is my default shell now. I have read How do I make Bash my default shell on Ubuntu? and chsh seams to be the preferred way to go. But when I type chsh /bin/bash I get this message: chsh: unknown user /bin/bash And if I just type /bin/bash the Bash shell seams to work fine. How to change it?

    Read the article

  • Shares not working on boot need to reinstall "File and Printer Sharing for Microsoft Networks" on server every morning to fix

    - by Neaox
    I had a problem a few days ago see question here: Can no longer access computer or network shares to my server from any other computers on the network The fix that I found does in fact work however when I boot my PC in the morning the shares are no longer working, to fix this I need to remote desktop into the server and re-install "File and Printer Sharing for Microsoft Networks" on the main adaptor. Doing this makes the shares work again, however it is anoying to have to do this each and every morning. On top of this my offline files are no longer available offline: I store my user profile on the server and had them selected to be "Always Available" however since this has happened they are no longer available offline and the option to make them available offline from the context menu is no longer available. Another problem, and I don't know if this is the cause or just a tell of a deeper issue but this server runs hyper-v, since these problems I can no longer remote desktop into the hyper-v client. Thanks for any help anyone can give.

    Read the article

  • Proxy - Some general questions

    - by user68802
    Is it possible to accomplish the following scenario with a proxy server? We are having one internet facing server that we want to put behind a proxy for some reasons. We want everything to work as before. When they do a request all connections will be forward to the internal server which will send back the information through the proxy. We want to be able to change to proxy to show an maintenance page whenever we are doing maintenance and change it back to forwarding traffic when we are done. We do also want to be able to keep forwarding all users that are using the sites but show an maintenance page for all new users for a time before showing the maintenance page for everyone in order to give the users some time to finish their work before kicking them out.

    Read the article

  • DVI problem with LG L204WT

    - by Holgerwa
    I have a LG L204WT which has VGA and DVI connectors. Until recently, I have used the monitor via VGA, but now I have upgraded the graphics board and switched to DVI. The problem is that each time after the computer starts, the monitor either doesn't show anything (black screen, but doesn't go into standby) or the picture is distorted. I then switch the monitor off and on several times, and at some point it starts correctly and then I can work as usual. This only happens on startup. It is one out of 3 monitors, the other two work fine. It seems to me as if the monitor is not able to synchronize on the DVI signal and either shows a distorted picture or nothing at all. There are no settings in the monitor setup that would change anything. What could be the problem?

    Read the article

  • How can I access blocked sites while in China?

    - by Samuurai
    I have some colleagues who need to go to China for work, however while they're there, they can't access a lot of sites. One of which is GMail (Google Apps), which they need for work. We have a UK based Ubuntu server, which I have root access for. What can I do for them? I thought about a Squid proxy, but it might rely on their hotel having port 8080 open, so, not ideal. Are there any workarounds or other solutions?

    Read the article

  • Linux cannot see Windows 7 partitions on install

    - by Nash0
    I've been trying to install Linux as a dual boot with Windows 7 on my Dell latitude e6510. It is currently running Windows 7 and I have used the MS disk tools to shrink the Win 7 NTFS partition to make room for Linux. The issue I'm having is that when I run Linux installers by boot from CD they see the entire hard drive as unallocated space. I have tried Ubuntu 10.10, Kbuntu 10.10 and Fedora 14 and they all have the same problem. I have also tried the Ubuntu "install in Windows" option and could not get it to work. EDIT: Booting Gparted 0.8.0 from a usb drive did not work. It reported the entire drive as unpartitioned.

    Read the article

  • Black screen after scheduled resume from sleep

    - by macbirdie
    I have a problem with my two Windows 7 setups at home. Whenever a scheduled task wakes up the machines from sleep, a black screen with a mouse cursor appears. If I move the mouse or press any key, desktop appears. The biggest problem is that during that black screen phase the display, sometimes even the PC too, doesn't go to stand by after a set period of time in power settings. If I get the desktop to show up, standby functions work fine. Display goes blank after a few minutes and as soon as the task doesn't need the PC, it goes back to sleep. It's frustrating that this "feature" is wasting energy and keeps lighting up the pc in the middle of the night for me to find it that way after I wake up. I found a handful of threads on the intertubes about the issue, but there were no answers in any of them. It sometimes happens in my XP machine at work as well.

    Read the article

  • Remote access and local access same hostname

    - by cpf
    Hi serverfault, I have a server in a clients network, seperated from theirs with a router/firewall, the intention is to have this server available through one hostname (example.com) My idea is to have (at least) a DNS server in the outside, to have outside (out of the clients' network) access the internal server. The problem would at that point be the internal client (PC A) My question: What would I have to do to make something like this work? Is it even possible or already done? The goal is to not have to change anything on either PC A or PC B, while both should access the same "internal server" while surfing to "example.com" Perhaps adding logic to the DNS server would work (Detect the external IP of internal client [PC A] is the same as the IP for example.com - Give the local IP as reply?) Anyhow: Thanks for helping me think on this!

    Read the article

  • Temporarily remove Group Policy in Windows 7

    - by rodey
    Using Windows 7 Professional I need to get access to the Folder Options screen. At work we primarily use Windows XP and are connected to a domain. Through Group Policy we disable access to the Folder Options. We use a tool called KillPol that temporarily removes Group Policy and let's me access the screen. This tool doesn't work on Windows 7 and I can't find a replacement. Any ideas on how I can get access to the Folder Options screen in Windows 7 when it is disabled through GP?

    Read the article

  • transparently set up Windows 7 as remote workstation

    - by Áxel
    Maybe is a very basic question, but I can't find the exact terms to Google for it and find the concrete answer to my doubt. Suppose we have several PCs in which individual employees work. One of them has an extremely powerful CPU, and it's very useful to use that computer to perform heavy computations, but go there and set up your task means its user has to stop working for a while. Is it possible to allow a secondary user account to remotly log in, for example via Remote Desktop, and work with a full user environment, while the main user keeps working under his user session? I've used remote desktop many times in the past, but it always blocked current user session, or even terminated it. Lots of thanks in advance guys.

    Read the article

  • iTunes high CPU usage

    - by Calm Storm
    I upgraded to iTunes 10.4.1 and use Windows 7 and my itunes library is not that large at all (say about 20gb) When I start iTunes the CPU goes between 60-80% and stays there for a long time. I see that the itunes.exe takes about 70% of CPU in Process Explorer and it spawns a SearchProtocolHost.exe every 2 mins or so which takes < 0.1% CPU. Other than that iTunes.exe is always at 70-90% and never lets me do anything else. Does someone have a suggestion? EDIT: I have tried reinstalling 10.4.1 completely deleting my library and starting with a plain installation and that does not work I have tried downgrading to 10.3.x and that does not work either :(

    Read the article

  • Cannot Login as root

    - by Josh Moore
    At my work we ship our product on pre-installed servers as a software/hardware package. We are using open SUSE 10.3 for the OS and we setup and we always log in with the root user to do maintenance on the box. Recently we just had box returned to us that the customer said the could not longer connect to the box through the network interface. So when I started to work on the box I run into the this problem: At the command prompt to login i type the user name "root" and hit enter. Then even before it asks me for a password I get "Login incorrect". I have never seen this behavior before and could not find any information about it online. Does anybody know what is going on? Thanks.

    Read the article

  • Switching to Chrome from IE

    - by Alan Parrish
    Hi, I work in IT at a school and we recently updated our database software, however the web access that the teachers use to do registration does not work too well with Internet Explorer 7 and we're thinking about switching to chrome (mostly due to me hating IE in general) but the problem we have is that users are unable to migrate their bookmarks over from IE due to account restrictions, is there any way to get this working? A bit of info about the system if it could be helpful; Almost all the client PCs are running Win XP SP2. My colleagues machines are running Windows 7, I use OS X Snow Leopard and most of out servers are running windows server 2003 (with the exception of 2 one on 2000 server and another on 2008 R2). The Active Directory Controller is running Windows Server 2003

    Read the article

  • How should a small team using multiple OS's deploy over github?

    - by Toby
    We have a small development team that have recently moved to using github to host our projects. The team consists of three developers, 2 on Windows and 1 on Mac. I am currently researching the best way to deploy applications to our Linux servers (dev and production). Capistrano running locally would be ideal but from what I read this won't work for Windows machines. It looks like the best way is to use a post-receive hook in github, I can see how this would work for auto deploying to dev, but I don't see how we could then deploy to live. I have found paid projects like http://www.deployhq.com/ but it feels like something that a quick bit of code should be able to do for free, I just can't seem to get myself pointed in the right direction! I was wondering what would be considered best practice for small team deployment involving multiple local OS's and github.

    Read the article

  • How to keep the Alt-Tab menu on the second monitor in Ubuntu?

    - by D Connors
    I'm running the latest Ubuntu version on my laptop, and every day when I arrive at my office I turn on the laptop and plug everything on. That includes a monitor that happens to be larger than the laptop's monitor. I always do my work on the desktop monitor, so I'd like for the Alt-Tab menu to always pop-up on this one. Instead, it's configured to pop-up whichever monitor the mouse is on (which makes sense, but doesn't work for me). Anybody know a way to change that?

    Read the article

  • How do I make control-click in Opera open new tabs?

    - by user21952-is-a-great-name
    Update: As of version 11.50, control-click opens the new tab in background, as desired. Yay! In Opera, as opposed to all other modern browsers I know, control-click does not open a new tab. Moreover, I couldn't find any good way to configure this behavior. The best available option seems to be this. However, it's so hacky that it won't work on HTTPS sites unless you enable user javascript there, which doesn't seem like a good idea. There also seem to be other proposed solutions, but none of them seem to work. I'd like this because my laptop has no middle mouse-button, and I'm a creature of habit. Do people have any ideas (for the latest version, 10.50)? I'd like ctrl-click to open a background tab, but I can do with foreground. Thank you!

    Read the article

  • How to get the Host value inside ~/.ssh/config

    - by iconoclast
    Within a ~/.ssh/config or ssh_config file, %h will give you the HostName value, but how do you get the Host ("alias") value? Why would I want to do that? Well, here's an example Host some_host_alias HostName 1.2.3.4 User my_user_name PasswordAuthentication no IdentityFile ~/.ssh/some_host_alias.rsa.id LocalCommand some_script.sh %h # <---- this is the critical line If I pass %h to the script, then it uses 1.2.3.4, which fails to give it all the options it needs to connect to that machine. I need to pass some_host_alias, but I can't find the % variable for that. (And: yes! I'm aware of the risk of recursion. That's solved inside the script.) UPDATE: Kenster pointed out that I could just hard-code the Host value as an argument to the script. Of course this will work in the example I gave, but it won't work if I'm using pattern matching for the Host.

    Read the article

  • IOException opening RFCOMM on openSUSE

    - by Chief A-G
    I have a permissions problem on openSUSE with Bluetooth /dev/rfcomm0. I've written a small test application which opens /dev/rfcomm0 and sends a request message and retrieves a response message. At first my problem was a permission denied error on /dev/rfcomm0 until I added the user account to the dialup users group. Now I get a System.IO.IOException Interrupted system call error whenever I run the app. I can sudo my application and it runs fine. I'm not sure how and which permissions to set to get this work work under my normal user account.

    Read the article

  • How can an RSA key be placed in a "non-folder"?

    - by Phil
    I lack a lot of knowledge regarding the subject so please excuse me if my question is very ignorant. I just started out with RSA key authentication to my test-server. It did not work at the beginning and I tried and tried and re-tried to no avail. Then I realised, I was creating "mkdir ~/.ssh" and then "mkdir ~/.ssh/authorized_keys" before placing the id_rsa.pub into the latter folder, which did not work. My question is: How can id_rsa.pub be placed in "authorized_keys", which is not created by "mkdir" therefore not a folder? Thank you.

    Read the article

  • Windows 7 + NVidia == 640x480 on external projectors

    - by ben2004uk
    Hello, I'm having a major problem and I really need your help! I have a MacBook Air which has the NVidia 9400M. The problem is, when I connect it to an external projector Windows 7 (and Vista) only allow me to connect at 640x480. I need it to use the laptop to present at conferences, as such I need it to work on a number of different projectors - at the moment I need to use OSX and VMWare but it's painfully slow and really doesn't work. At the moment, I'm considering buying a new laptop :( Is there any way to override the screen resolutions provided? I've saw some information around EDID? Thanks Ben

    Read the article

  • How to enable winhlp on Windows7 64bit?

    - by BGM
    Salvete! I just discovered that winhlp32.exe won't run on Windows7 64bit. I can't run the application, and I can't run hlp files either (but .chm files run fine). How do I make this work? I have downloaded the Microsoft fix here and restarted my computer, but to no avail. I can see the file winhlp32.exe in my c:\windows directory, but cannot run it. When I do run it, I get Windows' own "Help and Support" entitled, "Why can't I get Help from this program?" which sends me to the link above! How can I make it work?

    Read the article

  • Selectively routing traffic via ethernet or wifi, with proper DNS (Mac OS X 10.6)

    - by Dan
    When I'm at work, I access various intranet pages as well as the wider Internet through ethernet. However, the company LAN blocks some ports (e.g. Google Calendar). I can get to those through WiFi. So, I gave the Airport priority, and then using route add, I set up selective routing: all intranet traffic goes through the ethernet and everything else via WiFi: sudo route add 10.0.0.0/8 <intranet gateway>. However, there are a number of intranet sites that have their own DNS; i.e., hr.company.com only resolves on the intranet. The only way that I can get the DNS to work properly is to add the internal DNS server to the Airport DNS listing, however I fear that when I go elsewhere and forget, this will break things. What's the right way to get the DNS to resolve using this setup?

    Read the article

< Previous Page | 502 503 504 505 506 507 508 509 510 511 512 513  | Next Page >