Search Results

Search found 70655 results on 2827 pages for 'python time'.

Page 1840/2827 | < Previous Page | 1836 1837 1838 1839 1840 1841 1842 1843 1844 1845 1846 1847  | Next Page >

  • Recursively copy only new or changed files

    - by Niklas
    I'm sure this must have been asked and answered before, but I just can't find it right now... I have a Visual Studio post-build action that currently does a recursive copy (using xcopy) of an output folder to a different folder. This takes longer than I like and I'd like to only copy newly created and newer (changed) files each time (which xcopy doesn't seem to support). I cannot depend on any not-installed-by-default tool since the solution is used by different developers on different machines. What would a superuser do?

    Read the article

  • What is shrink_slab, and why is it showing up in /var/log/messages?

    - by Steve Kehlet
    I'm investigating some application issues on a Linux system and noticed this around the time of the most recent problem, but not close enough to be definitively related: Jun 7 17:51:49 localhost kernel: shrink_slab: nr=-155456000 Jun 7 17:51:49 localhost kernel: shrink_slab: nr=-157859400 Jun 7 17:51:49 localhost kernel: shrink_slab: nr=-157833400 Jun 8 06:20:23 localhost kernel: shrink_slab: nr=-284172800 What is this? Googling isn't helping, I get that it's something with VM, but why is it suddenly showing up in /var/log/messages? Edit: I take it it's showing this because the number of pages it could free ("nr") is negative. Which is presumably a bad thing. I found patches for newer kernels that display a more useful message to this effect. Still not sure what this means though, or what I need to do about it.

    Read the article

  • Best practice for administering a (hadoop) cluster

    - by Alex
    Dear all, I've recently been playing with Hadoop. I have a six node cluster up and running - with HDFS, and having run a number of MapRed jobs. So far, so good. However I'm now looking to do this more systematically and with a larger number of nodes. Our base system is Ubuntu and the current setup has been administered using apt (to install the correct java runtime) and ssh/scp (to propagate out the various conf files). This is clearly not scalable over time. Does anyone have any experience of good systems for administering (possibly slightly heterogenous: different disk sizes, different numbers of cpus on each node) hadoop clusters automagically? I would consider diskless boot - but imagine that with a large cluster, getting the cluster up and running might be bottle-necked on the machine serving the OS. Or some form of distributed debian apt to keep the machines native environment synchronised? And how do people successfully manage the conf files over a number of (potentially heterogenous) machines? Thanks very much in advance, Alex

    Read the article

  • Linux find/search root partition ONLY?

    - by ~sd-imi
    Say I need to do: find / -name somefile.txt and say root partition / is mounted on /dev/sda5; however, let's say I also have 250GB partitions (/dev/sda6, /dev/sda7) mounted in /media - AND another location that I cannot currently remember. Say, also, that I know the file I'm looking for is on /dev/sda5. Obviously, the above command will also descend in /media and that other directory which represent the big partitions, wasting time in looking for the file in the wrong place. Is there a way to instruct find (or other command) to search only / on /dev/sda5, and NOT to descend to directories if they are on different partitions ? Thanks, Cheers!

    Read the article

  • How to increase the disk cache of Windows 7

    - by Mark Christiaens
    Under Windows 7 (64 bit), I'm reading through 9000 moderately sized files. In total, there is more than 200 MB of data. Using Java (JDK 1.6.21) I'm iterating over the files. The first 1400 or so go at full speed but then speed drops off to 4ms per file. It turns out that the main cost is incurred simply by opening the files. I'm opening the files using new FileInputStream (and of course closing them in time to avoid file leaks). After some investigating, I see that Windows' disk cache is using only 100 MB or so of RAM although I have 8 GiB available. I've tried increasing the cache size using the CacheSet tool but any values I provide are considered out of range. I've also tried enabling the LargeSystemCache registry key but (after rebooting) the CacheSet tool still indicates I'm using 100 MB of cache (and doesn't increase during the test run). Does anybody have any suggestions to "encourage" Windows 7 to cache my 9000 files?

    Read the article

  • aptana/eclipse blocks other apps

    - by waquner
    Hi, I've got a fresh win7 installation, the only apps installed so far are aptana (beta 3)/eclipse and firefox + chrome (except antivirus and drivers) Every time I start aptana, chrome and firefox are blocked, so firefox can't load a webpage (or sometimes takes veeeery long) and chrome doesn't even take care of events (so for example hovering over a link doesn't change the cursor and a click does...nothing or typing in an url, press enter - nothing happens) First I thought, aptana downloads something and blocks my internet connection... but IE works well and also there are no signs in the taskmanager (cpu%, network or memory) When I close Aptana, all works as before. I tried updating Java, which didn't help... On my office-computer I use the same apps and it works perfect. Got any ideas? TIA, waquner

    Read the article

  • Mapped network drive missing from My Computer and Explorer

    - by matt wilkie
    On a Windows XP Pro SP3 machine one network drive refuses to show up in My Computer or Explorer. The missing drive letter is G:, if that matters. Other mappings work fine. Other profiles one the same machine have no problem seeing G:. I can access the G: just fine typing it into the address bar or in CMD shell. I've used TweakUI to toggle hide/show G: with no difference. TweakUI says G: should be visible. I've logged off,on between toggles to make sure the settings are taking effect. I've looked at reg key [HKEY_CURRENT_USER\Software\Microsoft\Windows\CurrentVersion\Policies\Explorer] and made sure it's zero'd.ref We've limped along with this broken setup for some time, just working around it, but some applications do not allow typing in a path when choosing a place to save files and it's reached the point where it's intolerable. So, anyone have any idea why XP won't show this drive letter? or how to fix it?

    Read the article

  • DELL Vostro 1520 bios crashed during update. now not giving even startup display

    - by Baahu
    my dell vostro 1520 crashed while i was installing drivers of windows 7 64bit. when i tried to update the bios it suddenly hanged. it remained in that state for a long time and was not getting off even by pressing and holding the power button. i removed the battery and after again installing the battery laptop is not giving any display of startup and all the caps lock lights are blinking. what the problem is this. i am worry about my laptop. please tell me the solution

    Read the article

  • Installing drivers and getting -> Error Code 28

    - by Adrov
    I've recently upgraded mainboard, without reinstalling OS, so I guess that's the issue. I really don't want to install OS at this moment. Issue is I can't install USB drivers, if I right-click uninstall, just installs same driver, which isn't working ofc. and giving error code 28. I fixed such issue once long long time ago, with editing registry, but I really can't remember what and where I have to do, so if anyone know please let me know, I'm also open to all other solutions to this issue. http://i.stack.imgur.com/AuBtB.png

    Read the article

  • Need recommendations for a hardy scanner that has a robust feeder tray

    - by JohnyD
    In the early days of our company all our information came in on paper and all of what we sold was on paper. Because of this we literally rent our an old bank vault to house the millions of sheets of paper that, some say, still contain relevant information. That being said, I'm looking into purchasing some hardware capable of scanning all these documents and converting them to pdf. Being new at this level of digitization I would like to ask for recommendations for accomplishing this task. Most of this material exists as separate bound studies/articles/etc. Someone would have to remove the bindings and be able to load many pages at a time and have the scanner feed them all through and convert them to a single pdf (single pdf per study/article/etc). If you have any recommendations I would very much appreciate hearing about them, thanks.

    Read the article

  • Changing location in Google Chrome when searching

    - by Alex
    I've recently moved to the Czech Republic from Scotland and I can't find a way to permanently stop Google from automatically defaulting back to google.cz all the time. I've checked to ensure that all my google accounts and cookie based settings (e.g. Advanced Search Options) are set to English but it's still clearly doing an IP address lookup and disregarding everything else. The default Search Engine for Google Chrome (and switches to google.cz automatically): {google:baseURL}search?{google:RLZ}{google:acceptedSuggestion}{google:originalQueryForSuggestion}sourceid=chrome&ie={inputEncoding}&q=%s I've tried hardcoding it to: http://www.google.com/search?{google:RLZ}{google:acceptedSuggestion}{google:originalQueryForSuggestion}sourceid=chrome&ie={inputEncoding}&q=%s this kind of works, but won't work for inline searching, i.e. I always have to press enter in order to get any results which is a bit annoying as I've gotten so used to AJAX style searching I can't have been the only one to get this issue? Any help is appreciated

    Read the article

  • options for physical architecture of rails site regarding caching server or cdn

    - by timpone
    I have a rails app that is set on a single server currently. On production, I force_ssl for everything. I am interested in using a caching server for images (I'm fine with css and js being served from origin for time being). Would nginx or varnish (which I have no experience with) be a better solution (for October 2012)? I'd imagine that it would be easy to switch these around while still on this single server architecture. Or would something like cloudfront (which I also have no experience with) make sense for hosting image files? I know this is a vague question but appreciate any current feedback. thx in advance

    Read the article

  • connection with the application object server could not be established

    - by user44048
    Hello All, I have installed Axapta server on my Axapta(Machine name) machine. When i am going to start the axapta at that time it give error like " connection with the application object server could not be established ". Basic all the steps that i have tested like Service start restart and connection configuration. port number problem. Still i am not able to resolve a problem. please can any one tell me the solution. It's urgent please help as soon as possible. Thank you.

    Read the article

  • Auto restart server if virtual memory is too low

    - by Sukhjinder Singh
    There are quite number of software running on my server: httpd, varnish, mysql, memcache, java.. Each of them is using a part of the virtual memory and varnish was configured to be allocated 3GB of memory to run. Due to high traffic load which is 100K, our server ran out of memory and oom-killer is invoked. We've to reboot the server. We have 8GB of Virtual Memory and due to some reason we cannot extend to larger memory. My question is - Is there any automated script, which will monitor how much virtual memory left and based upon certain criteria, lets say if 500MB left than restart the server automatically? I do know this is not the proper solution but we have to do it, otherwise we don't know when server will get OOM and by the time we know and restart the server, we lost our visiting users.

    Read the article

  • How can I get logrotate dateext to reflect the log date rather than the rotation date?

    - by Adam
    My Apache logrotate config looks like this: /var/log/http/*log { monthly dateext dateformat .%Y.%m [... rest stripped for brevity ...] } This works great, except that the date on the rotated filename is one period later than the period actually covered by the logs, for example error_log.2012.09 covers 2012-08-01 to 2012-08-31. I realize that there are other options for Apache (eg. cronolog), but I have a bunch of other logs that I also need to rotate, and logrotate is really exactly what I need apart from this one issue. Is there a way to get logrotate to use a date offset -- or, even better, figure out the previous time period -- when generating the rotated filename?

    Read the article

  • I can't connect to my WAMP server from a client on the same network

    - by F4r-20
    Usually I would manually set up Apache and PHP but this time I thought I would go for the WAMP alternative. I've set up WAMP and it's working perfectly fine on the host machine. If I try and access the WAMP server from another machine on the same network (same subnet too: 192.168.1.*) I get this message: Forbidden You don't have permission to access / on this server. I have been online and gone through steps. I have changed Allow from 127.0.0.1 to Allow from all and it's still not working. Is there anything else I can try? Thanks.

    Read the article

  • Linux virtual disk stripping or multi-path samba share?

    - by wachpwnski
    I am trying to build a file storage box for media. It needs to span two or more directories or partitions as one share. There are a few solutions but reasons why I want to avoid them, among these are: Using LVM2 for stripping. I don't really have the resources to back up everything on the volumes incase one HDD goes south. I would end up loosing everything. Maybe there is a better option for this to prevent data loss with hot swappable drives or some kind of raid. Using symbolic links in the share. This will get tedious every time a new sub-directory is added. Is there some kind of software raid I can use to merge two directories virtually? I am aware of the issue where /dev/hda1/media/file.1 and /dev/hdb1/media/file.1 both exist. But I'm sure there are some creative solutions for this.

    Read the article

  • Virtualbox PUEL Interpretation

    - by modernzombie
    Sorry if this seems like a lame question but I want to be sure before making a decision. The Virtualbox PUEL license says “Personal Use” requires that you use the Product on the same Host Computer where you installed it yourself and that no more than one client connect to that Host Computer at a time for the purpose of displaying Guest Computers remotely. I take this to mean that if I want to setup a development server (web server) that's only used by me to do my work this falls under personal use. But if I make this server available for clients to connect to the websites to view my progress this is no longer personal use also meaning that using Vbox to run a production web server is also against the license. Again sorry if this is a dumb question but I find it hard to follow the wording used in licenses. I know I could go with OSE but I have not looked into VNC versus RDP yet. Thanks.

    Read the article

  • How to transer/execute a script on remote unix server from Linux machine

    - by Jagadeesh
    I am trying to deploy an executable and execute it on remote Unix machine(Linux/Solaris) from Linux without entering the password manually. I tried scp and also SSH key-gen utility to interact with remote server but in either way i couldn't avoid providing the password manually. Since I need to run this command/utility from Java code, I should completely avoid prompting for the password at run time. I have gone through many topics in google but nothing has been materialized. Your help would really be appreciated if i could proceed further on this issue. Thanks in advance. -Jagadeesh

    Read the article

  • Missing files when Windows 7 returns from hibernate w/ dual boot

    - by Arthur N
    I have a dual-boot setup with Ubuntu (lucid) and Windows 7. I have the Windows file system shared on Ubuntu through Samba. Occasionally, I am working on Windows and my machine will go into hibernate (i.e. when the battery level is critical). By default, my GRUB settings boot me into Ubuntu. So when I get back to my PC, sometimes I just hop into Ubuntu instead of going back to Windows. However, if I write any files to the Windows file system during that Ubuntu session, the next time I do go back to Windows (which resumes from hibernate), those files are missing. Obviously, the state of the actual file system and the hibernate snapshot become out of sync, and Windows chooses the hibernate snapshot, overriding any changes I may have made thru Ubuntu. For now, I've disabled the hibernate option in the Windows power settings, but is there any utility I can use to get back some of those missing files?

    Read the article

  • Download web server structure with empty files

    - by golimar
    I want to make a mirror of a Web server, but downloading the actual files will take too long. So I thought of having just the directory and file structure, and when I need the actual contents of the file, I can download just that file. I have tried wget --spider URL and in a short time it has created in my local disk the directory structure with no files. But I've checked all of wget's or curl's switches and there is nothing like what I need. Can this be done with wget, curl or any other tool?

    Read the article

  • How do I force Windows to play sound through the speakers only when a USB headset isn't connected?

    - by Phoexo
    I'm using a speaker set connected through the green audio jack and a headset which I connect through USB. My problem is that every time I connect/disconnect my headset, I have to go through a lot of settings/restart some programs to make the sound go through the speakers again. What I want is to have audio play through the headset when it's connected, but if I disconnect the headset, I want the audio to automatically play through the speakers. For example, if I connect/disconnect the headset while listening to music, I have to restart the application to make the music play through the correct speaker/headset, and it shouldn't be that inconvenient. (I found this somewhat relevant topic, but the problem is that it doesn't really give an answer. (Also, it is 2 years old.))

    Read the article

  • One subdomain is not working

    - by BFTrick
    Hello there, My main domain works just fine - www.example.com and a subdomain set up by another developer works as well - sub1.example.com. But when I try to set another subdomain up I go through the process everything seems to work. The software creates the default files where the subdomain files should go. But when I try to browse there it doesn't work. My host uses Plesk to do all of the hosting stuff. What do you think the problem is? I doubt it is some sort of cache issue because I had problems on my phone which I tried after problems on the pc. Maybe for some reason Plesk needs time to set this up? I have used Cpanel before and that works instantly.

    Read the article

  • Renaming Files & Folders in Windows 7

    - by Mart
    This has been happening to me for some time now. I'm currently running Windows 7 RTM. When I try to rename a new folder that is created (by copying usually), I usually get a "Folder is in use" error. I would have to wait a while before I would be able to modify the folder name. Files behave the same way too. I am thinking that this is because I am sharing the folder, or library, via Homegroup Sharing. Is anyone who is running Windows 7 RTM experience this as well? EDIT 1: I'm using Windows 7 Ultimate.

    Read the article

  • How could I compress a folder into splitted archives (individual ZIPs)?

    - by Shiki
    I have to compress folders into ZIP packages. But the size is limited, only a ~10-15mb is allowed to used per package. Every major application comes with the "Split archive to..." option, which does what I want... except I can't uncompress them one-by-one. (You need them all, and then use the .7z, .rar, .zip file to uncompress.) Here is an example. FolderX is 35 mb. That makes 4 packages, 4 zip files. The normal split function would give me: folderx.zip, folderx.zip.001, folderx.zip.002, folderx.zip.003 What I would really need is: folderx_1.zip, folderx_2.zip, folderx_3.zip, folderx_4.zip (Individually uncompressable files/packs.) I can code this down into an app, but it's a waste of time if such a utility already exists.

    Read the article

< Previous Page | 1836 1837 1838 1839 1840 1841 1842 1843 1844 1845 1846 1847  | Next Page >