Search Results

Search found 9091 results on 364 pages for 'thinking'.

Page 140/364 | < Previous Page | 136 137 138 139 140 141 142 143 144 145 146 147  | Next Page >

  • Renaming Files & Folders in Windows 7

    - by Mart
    This has been happening to me for some time now. I'm currently running Windows 7 RTM. When I try to rename a new folder that is created (by copying usually), I usually get a "Folder is in use" error. I would have to wait a while before I would be able to modify the folder name. Files behave the same way too. I am thinking that this is because I am sharing the folder, or library, via Homegroup Sharing. Is anyone who is running Windows 7 RTM experience this as well? EDIT 1: I'm using Windows 7 Ultimate.

    Read the article

  • Drobo Pro or SAN for backups

    - by PHLiGHT
    I have multiple offices and was thinking of using two Drobo Pros as a storage location for my backups, with one at each site being synced. It seems like on the surface its a cheap alternative to a SAN. Because it will just be used for storing backups I don't think that performance is as big of an issue. Related question. How do you sync your backups from your main location to a remote site. I am using Backup Exec and wondered if I should use a feature in the program or use something like Delta Copy.

    Read the article

  • Is there a difference in page fault rates between CPU bound and I/O bound processes?

    - by user198864
    I was thinking, should there be any difference in expectation of the page fault rate on CPU-bound vs I/O bound processes? At first I thought maybe we could, since CPU-bound processes would likely be using more memory accesses per time quantum, so I expect it would move from locality to locality faster. At the same time, the CPU-bound process is probably given a larger working set... but this doesn't affect the fault overhead as it hits a new locality IF this wasn't pre-paged in. Is there actually any real difference in the page fault rates or am I just musing about something nonexistent? And if there is, how would it impact a real-world OS like linux?

    Read the article

  • How to create batch file that delete all the folders named `bin` or `obj` recursively?

    - by Nam Gi VU
    I have the need of deleting all bin & obj folders in a folder on my PC. So, I'm thinking of a batch file to do that but I'm not famaliar with batching file in Windows. Please help. [Edit] After discussion with user DMA57361, I got to the current solution (still having problem though, see our comments): Create a .bat file, and paste the below command: start for /d /r . %%d in (bin,obj) do @if exist "%%d" rd /s/q "%%d" OR start for /d /r . %%d in (bin,obj) do @if exist "%%d" rd /s "%%d" @DMA57361: When I run your script, I get the below error. Any idea?

    Read the article

  • Laptop crashes when connecting to external harddisk

    - by Gnot
    I recently had a problem with my laptop. when I booted up the machine, I would get a SMART failure error message and when I pressed F1 to continue, it would take a very long time to boot and it would come back to the same error message again. Thinking that my hard disk was dying, I bought a new hard disk and installed on my laptop and so now my laptop is alright. However I need to recover data from that old hard disk, so I bought an external hard disk case and placed the old hard disk onto the case and connected to my laptop with USB. The first few times when I connected, I could see the files from the old hard disk and managed to copy some files over although it took extremely long to transfer. But now whenever I connect to the old hard disk, after a few minutes, my laptop will crash and re-boot. Do you think my old hard disk is dead beyond repair? Or you can offer some help here? Any assistance would be appreciated!

    Read the article

  • How to kill all screens that has been up longer then 4 weeks?

    - by Darkmage
    Im creating a script that i am executing every night at 03.00 that will kill all screens that has been running longer than 3 weeks. anyone done anything similar that can help? If you got a script or suggestion to a better method please help by posting :) I was thinking maybe somthing like this. First do a dump to textfile ps -U username -ef | grep SCREEN dump.txt then do a loop running through all lines of dump.txt with a regex and putting pid of the prosseses with STIME 3weeksago in a array. then do a kill loop on the array result.

    Read the article

  • Is there a way to create a copy-on-write copy of a directory?

    - by BCS
    I'm thinking of a situation where I would have something that creates a copy of a directory, tweaks a few files, and then does some processing on the result. This wold be done fairly often, maybe a few dozen times a day. (The exact use case is testing patch submissions; dupe the code, patch it, build/test/report/etc.) What I'm looking for could be done by creating a new directory structure and populating it with hard links from the origonal. However this only works if all the tools you use delete and recreate files rather than edit them in place. Is there a way to have the file system do copy-on-write for a file? Note: I'm aware that many FSs use COW at a block level (all updates are done via writes to new blocks) but this is not what I want.

    Read the article

  • Windows 7 / Internet Explorer 8

    - by Rene
    I am a shop owner at zazzle.com. About six weeks ago, when my computer was running on Windows XP/IE7, my sites, as well as zazzle's homepages went out on me. I can only see part of each page. Since that time, I have a new computer running Windows 7/IE8, thinking that would solve the issue. It did not. Zazzle's emails told me to download Firefox and/or download Internet Explorer 7. I tried Firefox and was getting a different problem at the zazzle site. Now I was getting only the 'view source' pages on zazzle's homepages and my own shop sites as well. Question: Can I download IE 7 onto my IE 8 computer? Can this be done without loading that compilation of internet explorer 1 through 8? What do you think is the best solution to this problem?

    Read the article

  • Suitable companion monitors for HP LP3065 30" primary monitor

    - by Bittercoder
    I have an HP LP3065 30" monitor - now I want to add a couple of companion screens to either side of it - I've heard to people suggesting 21" monitors in portrait mode - but it seems difficult to find a suitable panel these day, as 22" wide-screen monitors have become common, and I can't seem to find anything with a suitable 1600 pixel width. Can anyone recommend a suitable monitor, and suggest what criteria are best for picking a suitable monitor - what's important to consider - I'm thinking things like: Exact resolution match? Viewing angle? Small/thin bezel?

    Read the article

  • Backup folder on sometimes attached external usb harddrive

    - by ctrler
    My girlfriend no longer has space her laptops drive to store her photos. The drive she has now is 750GB, so to go to a bigger drive would be expensive, as there isn't many 1.5tb 2.5 inch 9.5mm hdd on the market (as of now, there is only one). Because of that, I am thinking of moving her pictures to a cheap external usb hdd. As of now, I'm automatically backing up her important folders (My Documents, Pictures, etc.) using Windows 7 default backup software to a network drive. My problem is that I don't know of a good solution to automatically backup a folder residing on an usb disk. The usb disk won't be attached to the computer all the time, so I can't just treat it as a normal backup folder. Sometimes the backup would run and the folder would not be there! Anyone knows any software or methodology to backup folders on external usb hard drives that are not always present? Thanks

    Read the article

  • What is a good listserv replacement?

    - by David Smith
    A board I'm on has about 700 - 1300 members being served from a listserv maintained by a company one of our board members is on. This restricts the ability of other board members to use the listserv unless they are employees of the company. They want to move to a system where anyone can administer the list. I'm thinking of switching to Constant Contact or Google Groups. We want something that allows list members to manage their own contact preferences and enables anyone authorized to send emails to the list.

    Read the article

  • How can I "shadow" the filesystem on Linux?

    - by happy_emi
    On a Linux environment sometimes I need to run a script as root which will add/modify serveral files on my fs. Basically I'd like to know exactly which files are modified and how WITHOUT opening the script and trying to guess the code. I was thinking about using something like unionfs: the main fs would be accessible in readonly mode and all changes are written on a file used as a partition and "mounted" in write mode. Are there other ways to achieve the same goal (i.e. other than unionfs)?

    Read the article

  • Does it make a difference to read from a file instead of from MySQL?

    - by Joe Huang
    My web server currently is quite loaded. And I have a PHP file that is accessed very often remotely. The PHP file basically makes a MySQL query and returns a JSON formatted string. I am thinking to use a Cron job to write the necessary data into a file every 15 mins, so the PHP file doesn't make a MySQL query, instead it reads from the file. Does it make a difference? I mean to alleviate the server loading (CPU/MySQL) a bit?

    Read the article

  • Move /var directories to to /mnt on an EC2 instance

    - by Geoff Lanotte
    I am trying to work on a standard configuration for a set of EC2 instances running ubuntu 12.04. These servers are going to be primarily web servers for a Ruby on Rails application. When you configure a new large instance, you are given a primary of 8GB and then ephemeral storage of 400 GB that is mounted to /mnt. It seems logical to me to move some directories that have a potential for growth off to the /mnt directory, I was specifically thinking of /var/www and /var/log. My question is two-fold: Is this a good idea or are there pitfalls that I cannot see? If this is a good idea, how should I go about configuring this. I do have the ability to configure new instances and down our old instances. My concern is over long term, doing this in such a way that it prevents downtime. I am a developer with some experience in devops, but mounting drives is something I have not faced before, so explicit directions would be greatly appreciated.

    Read the article

  • New Servers Active Directory and Exchange

    - by user3164638
    I have 3 Dell PowerEdge server, each with 2 quad-core processors. I am going to bring this office out of the stone-age network (P2P, share files on a flash drive, emails through Google, etc) and set up Active Directory and Exchange 2013. Our needs are not that great at the moment - our staff consists of approximately 40 people, and our network may eventually be managed by an external company. We need only one domain for our emails (though we may serve emails for a few other partners domains as well). I was thinking of setting something up like this: Server 1: Primary DC. Active Directory and Exchange on separate virtual machines. Server 2: Redundant of server 1. Server 3: Shared resources, storage, backups, etc. How would you utilize 3 servers for an Active Directory / Exchange setup for a small/medium office? We do have plans to grow, so my solution must be scalable, though I'm not sure that I want to split permissions, though I'd consider it if that was something that could be changed on down the road.

    Read the article

  • Vagrant sahara plugin - multiple snapshots

    - by BazZy
    How I can make more than one snapshot when in sahara sandbox mode? Or, can I see list of all commits I've ever made and rollback to any? Why do I need all this? I just want to setup initial VM. After that I want to compile number of packages from sources, and this takes pretty long time. So right now I thinking of this order: Setup initial ubuntu 12.04 vagrant box Snapshot this state Compile sources, install system wide rbenv (it make compilation also) Snapshot second state Start all my infrastructure experiments Rollback to any of my previous states or commit third state

    Read the article

  • Conditional cPanel Forwarder

    - by Wireblue
    We have many clients on a cPanel server, some of which also have SSL certificates setup. Each year when the renewals are sent to webmaster@ their domain we would like a copy of those emails so we can install the SSL certifciate for them and issue invoices etc. So I'm wondering if there is a way we can selectively or conditionally forward certain emails if they match certain rules? (in cPanel) I'm thinking: If an email is sent to "webmaster@domain" and the subject contains "ssl", then forward to "[email protected]". Any ideas? Thanks in advance, Wireblue

    Read the article

  • tricks for speeding up tar while tarring up a huge directory of little files?

    - by Trevor Harrison
    I'm trying to tar up a directory that has about 3M tiny files in it. Tar is chugging along, but I'm thinking its going to take longer than I can wait. I'm wondering if telling tar to not store metadata (owner, group, perms) would reduce the churn on reading and re-reading this huge directory and maybe speed things up, and if there is a tar switch that does this. My initial perusal of the man page only gets me something like --no-xattrs, which looks like a start, but I was hoping someone had some specific knowledge.

    Read the article

  • two shops network

    - by edward
    okay so, I just opened up two shops in my hometown. The two stores is about 6 blocks apart, connecting them by wire is not really feasible in cost wise. What kind of network topology should I use for my small shops, there will be 5 computers, one is the sales computer ,another 4 as mentioned is the guest computer. I want the sales and guest computers network to be seperated. Both shops have same computers. The guest computers serves up simple website that has my shop catalog on it, I'm thinking of using a web server. So, how am i suppose to setup these networks, im planning to add in more computers in the future. Is it I need to station a single server at a shop, and all the computer connected to it? or is there any more effective methods? I'm no networking expert, would love to hear some advice.

    Read the article

  • Windows Command queue?

    - by Stefano
    i'm thinking if does exist some kind of software that can put in a queue a bunch of windows commands... for example i can say to first copy some file somewhere, then rename those, then delete the old files, then edit one of them etc.... without waiting the effective execution of any of those passages.... this could be useful when copying big files that take a lot and i don't want to sit in front of the computer keeping the eyes on the progress bar... does exist anything like this?

    Read the article

  • X11 forwarding through SSH

    - by martinjlowm
    I have been playing around with X11 forwarding the past few hours and so far I've managed to forward my desktop pc's X Server to my laptop, using X11VNC as server and X2VNC as client. X2VNC uses Xinerama to provide a dual-screen-like behavior between my laptop and my desktop pc. It's actually really great! I know that most Linux systems run Xorg and desktop environments on TTY7. Therefore I was thinking, is it in any way possible to have the VNC-tunnel tied to it's own TTY? It would be great to be able to switch forth and back between two TTY's in order to choose which machine to manage. And I would like this approach more than using Xinerama or a GUI.

    Read the article

  • Sync Linux to Windows 2003/2008 Natively

    - by user26753
    Without using any external packages for Windows, can Linux synchronise it's clock to a Windows 03/08 Server natively? I've tried it using various NTP packages for Windows but would like to use just Windows software for this. EDIT: I've tried the below however it doesn't work. I've put TimeSource in /etc/hosts, stated server TimeSource in /etc/ntp.conf however when I do a service ntpd start it doesn't sync (it's 3 minutes out at the minute). I then stopped the NTPD and did a ntpdate TimeSource, thinking it would sync and then I'd start the NTPD once it's got the time, and it says: no server suitable for synchronization found. Though I can ping it. Any thoughts?

    Read the article

  • Hardware specs for web cache

    - by Raj
    I am looking for recommendations for hardware specs for a server that needs to be a web cache for a user population of about 2,000 concurrent connections. The clients are viewing segmented HTTP video in bitrates ranging from 150kbps to 2mbps. Most video is "live" meaning segments of 2-10 secs each, of which 100 or so are maintained at a time. There are also some pre-recorded fixed length videos. How would I go about doing the provisioning calculation for such a server: What kind of HDD (SSD?), how many NICs how much RAM etc? I am thinking of using Varnish on Linux, all the RAM I can get my hands on, 2 CPUs with 6-8 cores each.

    Read the article

  • Is there a good, free way to fix broken/corrupt .wmv files?

    - by chbtn
    I've recovered some files from an hdd that weren't supposed to be deleted in the first place, but they have seeking problems/crash the players. Since they have the right size, I'm thinking it might be a problem of corrupt index/header, so I'm trying to find a way to fix them. It's easy to find examples on how to fix corrupt .avi files with mencoder, but .wmv seems trickier. Also, I realize there might not be a way to fix these files, but I figure I might as well as try. As far as players go, I've tried opening it with vlc/mplayer/windows media player. I can use anything on Windows XP/7 and Ubuntu, as long as it's free. Since the files are 200mb+ and there are quite a few, I don't think trial software would work.

    Read the article

  • Whats the best way to deal with backups for my php/mysql application

    - by spirytus
    I'm creating php application for my client and now thinking what would be the best way to do backups, automatically if possible? I don't have much experience in this area and in case something goes wrong, or if I need to migrate, I would like to have fast way of getting it all back online. I understand "something goes wrong" is a very wide term, but lets say that someone hacks my site and wipes out database and all the files. My app. is written in php/mysql and I got access to cpanel (hosted with justhost.com if that makes any difference :). I used Joomla and it has JoomlaPack that does complete backup almost automatically and in case site fails, its easy to revert, or migrate if necessary. Is there anything like that for my configuration that would make reverting/migration, easy?

    Read the article

< Previous Page | 136 137 138 139 140 141 142 143 144 145 146 147  | Next Page >