Search Results

Search found 25196 results on 1008 pages for 'hard drive cache'.

Page 279/1008 | < Previous Page | 275 276 277 278 279 280 281 282 283 284 285 286  | Next Page >

  • What conditions cause a web browser to display the "Page Expired" message?

    - by RichardHowells
    I assumed that sending a page out with an expiry (a la)... Response.Cache.SetExpires(System.DateTime.Now.AddSeconds(5)); Response.Cache.SetCacheability(HttpCacheability.Public); Response.Cache.SetValidUntilExpires(true); ...would mean that if the user hit the back button they would see the expired page. It does not seem to work that way. I found that using the back button would simply show the old (IMO expired) page. In fact in experimenting with different combinations of caching/not caching and expiry times I NEVER managed to get the "Page Expired" message out of the browser. What conditions provoke that message? Server Environment ASP .Net. I've only tested in IE8 - I'm assuming that other browsers are consistent here.

    Read the article

  • How much faster are register based architectures than stack architectures?

    - by drozzy
    Studying compilers course, I am left wondering why use registers at all. It is often the case that the caller or callee must save the register value and then restore it. In a way they always end up using the stack anyway. Is creating additional complexity by using registers really worth it? Excuse my ignorance. Update: Please, I know that registers are faster than RAM and other types of cache. My main concern is that one has to "save" the value that is in the register and the "restore" it to. In both cases we are accessing some kind of cache. Would it not be better to use cache in the first place?

    Read the article

  • Save NSCache Contents to Disk

    - by Cory Imdieke
    I'm writing an app that needs to keep an in-memory cache of a bunch of objects, but that doesn't get out of hand so I'm planning on using NSCache to store it all. Looks like it will take care of purging and such for me, which is fantastic. I'd also like to persist the cache between launches, so I need to write the cache data to disk. Is there an easy way to save the NSCache contents to a plist or something? Are there perhaps better ways to accomplish this using something other than NSCache? This app will be on the iPhone, so I'll need only classes that are available in iOS 4+ and not just OS X. Thanks!

    Read the article

  • Will lock() statement block all threads in the proccess/appdomain?

    - by MikeJ
    Maybe the question sounds silly, but I don't understand 'something about threads and locking and I would like to get a confirmation (here's why I ask). So, if I have 10 servers and 10 request in the same time come to each server, that's 100 request across the farm. Without locking, thats 100 request to the database. If I do something like this: private static readonly object myLockHolder = new object(); if (Cache[key] == null) { lock(myLockHolder) { if (Cache[key] == null) { Cache[key] = LengthyDatabaseCall(); } } } How many database requests will I do? 10? 100? Or as much as I have threads?

    Read the article

  • Caching images with different query strings (S3 signed urls)

    - by Brendan Long
    I'm trying to figure out if I can get browsers to cache images with signed urls. What I want is to generate a new signed url for every request (same image, but with an updated signature), but have the browser not re-download it every time. So, assuming the cache-related headers are set correctly, and all of the URL is the same except for the query string, is there any way to make the browser cache it? The urls would look something like: http://example.s3.amazonaws.com/magic.jpg?WSAccessKeyId=stuff&Signature=stuff&Expires=1276297463 http://example.s3.amazonaws.com/magic.jpg?WSAccessKeyId=stuff&Signature=stuff&Expires=1276297500 We plan to set the e-tags to be an md5sum, so will it at least figure out it's the same image at that point? My other option is to keep track of when last gave out a url, then start giving out new ones slightly before the old ones expire, but I'd prefer not to deal with session info.

    Read the article

  • What is the cost of memory access?

    - by Jurily
    We like to think that a memory access is fast and constant, but on modern architectures/OSes, that's not necessarily true. Consider the following C code: int i = 34; int *p = &i; // do something that may or may not involve i and p {...} // 3 days later: *p = 643; What is the estimated cost of this last assignment in CPU instructions, if i is in L1 cache, i is in L2 cache, i is in L3 cache, i is in RAM proper, i is paged out to an SSD disk, i is paged out to a traditional disk? Where else can i be? Of course the numbers are not absolute, but I'm only interested in orders of magnitude. I tried searching the webs, but Google did not bless me this time.

    Read the article

  • What is the best way to prevent a webpage coded completely in HTML (no PHP or any server side langua

    - by racl101
    I thought by placing these two lines that it would prevent the page from being cached but this is not the case between the head tags: <meta http-equiv="pragma" content="no-cache" /> <meta http-equiv="cache-control" content="no-cache" /> As already mentioned this is on server that doesn't support PHP (or any server side language, well, that's not true, it has cgi-bin folder that supports perl scripts but I don't know perl and have no time to engage in it.) What else could I do to prevent the page from caching? It caches on Firefox and IE (and not so much in Chrome or Safari.)

    Read the article

  • Dynamic control

    - by Geetha
    Hi all, I am creating dynamic label and textbox based on the number of values from the database for the selected item of the dropdownlist. Then the dynamic labels will have the names and the text box with the values. To retain the values of these controls im using Page_init event. So im using cache to hold the selectteditem from the dropdownlist. Problem: The process is going fine. But if i try to refresh the page no items are selected in the dropdown list but the cache is not getting clear so using this cache value dynamic control are creating. Geetha

    Read the article

  • Recommand a Perl module to persist a large object for re-use between runs?

    - by Alnitak
    I've got a large XML file, which takes 40+ seconds to parse with XML::Simple. I'd like to be able to cache the resulting parsed object so that on the next run I can just retrieve the parsed object and not reparse the whole file. I've looked at using Data::Dumper but the documentation is a bit lacking on how to store and retrieve its output from disk files. Other classes I've looked at (e.g. Cache::Cache) appear designed for storage of many small objects, not a single large one. Can anyone recommend a module designed for this?

    Read the article

  • how to force browser to re-request page after history.back()

    - by yellowred
    Hi, I've got two scripts: 1.php and 1.php. Here they are: 1.php <?php header('Pragma: no-cache'); header('Cache-Control: max-age=1; no-cache'); header('Expires: Tue, 1 May 1985 01:10:00 GMT'); header('ETag: "'.md5(rand(1, 1000)).'"'); print date('H:i:s'); ?> <a href="2.php">pay</a> 2.php <a href="javascript:history.back()">back</a> Visitor lands on 1.php and then goes to 2.php. I want browser to re-request 1.php after vistor's click on "back" link. On current state it doesn't work. How it can be managed?

    Read the article

  • Best choice for a personal "online backup" in Europe

    - by marc_s
    I'm looking for an online backup solution for personal use - besides all the usual requirements (like not too expensive, since it's for personal use), I'd like to add two requirements to it: data center should be in Europe (I don't want my personal data stored in the US, when the next crazed president comes along and wants to confiscate and rifle through everybody's files.....) the online backup store should be accessible through a drive letter in cmd.exe So far, I've looked at a few services, but none have totally convinced me: Dropbox is looking ok, but they insist on creating a silly "My Dropbox" directory in my data path - and there's no way I can choose that name. Sorry - "My everything" is for dummies - I don't like that, I like to name my files and folders according to my liking LiveDrive is OK, too - they offer European storage, drive letter and all - but those drive letters are only available in the Windows Explorer - and not on the cmd.exe command line :-( and since I do 99% of my work on the command line, this is a major drawback..... Any other services I haven't looked at worth checking out? Marc

    Read the article

  • Can't re-install OS on Asus Eee PC 1018P

    - by cornerback84
    My friend has an Asus Eee PC 1018P. It has no CD/DVD drive (neither does he have a USB CD/DVD drive). The OS wasn't working fine, so we decided to restore the system using the provided OS backup from HD. But mid way through the installation it was interrupted and the computer was restarted (not a hardware or software issue – we did it). Now we cannot run the backup and we also tried to install Windows 7 through USB, but as soon as we start to install the OS, it says that some device driver is missing. It turns out that the OS needs a USB device driver to continue. It has USB 3.0 – maybe that's why the OS needs the driver. I tried disabling 3.0 and enabling 2.0 but then it does not boot from USB drive for some reason. We are stuck with this. The backup doesn't run and when booting from USB, it says that it needs a device driver. Anyone has any idea what we could do?

    Read the article

  • Using "offline files sync" to sync with a local resource [closed]

    - by Kije
    Possible Duplicate: Which is the best application to Sync two folders? I have been trying get my machine (XP-Pro SP3) to sync files to a local USB drive in the same way as I can with mapped network drives. I particularly want to the sync to happen automatically when the USB is connected - in the same way that Off-line files will sync when the network drive comes on line. I can get a folder on the USB mounted as a network drive, but cannot get XP to offer the off-line files functionality. I have tried MS's SyncToy - it works as advertised, and will do scheduled and ad-hoc sync's, but does not seem to offer automatic syncs. I suspect I could do this by plugging my USB into another machine on my network - but that seems heavy handed. All insight appreciated - If you know this cannot be done please say so.

    Read the article

  • Proliant ML350 Won't Load Windows

    - by Mike
    I have a HP Proliant ML350 G4 running Windows 2003 Server Std Edition that will not load Windows on boot. It gave me a BSOD while running a recent Windows update and has not been able to boot since. The BSOD was a generic error, "Hardware Fault. Contact vendor." I am not getting any errors when the machine boots--it just hangs with a black screen when Windows starts to load. The machine will not boot into safe mode either--it gets stuck after beginning to load drivers. When trying to load the Windows recovery console I get the message that Windows cannot find any hard disks on the machine. This, despite having to load the driver for my storage controller. The HP Smart Start diagnostics finds no problems with my hard disks nor anything else. What should I try next? Mainly, I just want to be able to pull the data off my hard disks.

    Read the article

  • What’s the password for my FileVault 2 boot volume?

    - by cbowns
    Apple’s support document for FileVault 2 (a.k.a. “full disk encryption” or “FDE”) has lots of information about enabling FDE and what it means for booting the machine. However, it doesn’t cover one very important thing I’m trying to do: mount the drive in the Recovery HD environment to reinstall OS X on it. The Recovery HD environment asks me for the volume passphrase so it can mount my drive and try to install OS X onto it. If this were an external drive which I’d manually enabled FDE on with diskutil, or an external Time Machine volume, I’d know it because it makes you pick one (just like a regular login password), but FileVault 2 never asked me for a volume passphrase (I assume it selects one behind the scenes). I’ve tried my main user’s password, but that doesn’t work, and neither does the recovery key set for the volume. Keychain Access doesn’t have anything that I could find. How do I unlock this volume?

    Read the article

  • Hyper-V File Server Clustering - at my wit’s end

    - by René Kåbis
    I am at my wit’s end with File Server clustering under Hyper-V. I am hoping that someone might be able to help me figure out this Gordian Knot of a technology that seems to have dead ends (like forcing cluster VMs to use iSCSI drives where normally-attached VHDX drives could suffice) where logic and reason would normally provide a logical solution. My hardware: I will be running three servers (in the end), but right now everything is taking place on one server. One of the secondary servers will exist purely as a witness/quorum, and another slightly more powerful one will be acting as an emergency backup (with additional storage, just not redundant) to hold the secondary AD VM and the other halves of a set of clustered VMs: the SQL VM and the file system VM. Please note, these each are the depreciated nodes of a cluster, the main nodes will be on the most powerful first machine. My heavy lifter is a machine that also contains all of the truly redundant storage on the network. If this gives anyone the heebie-geebies, too bad. It has a 6TB (usable) RAID-10 array, and will (in the end) hold the primary nodes of both aforementioned clusters, but is right now holding all VMs. This is, right now: DC01, DC02, SQL01, SQL02, FS01 & FS02. Eventually, I will be adding additional VMs to handle Exchange, Sharepoint and Lync, but only to this main server (the secondary server won't be able to handle more than three or four VMs, so why burden it? The AD, SQL & FS VMs are the most critical for the business). If anyone is now saying, “wait, what about a SAN or a NAS for the file servers?”, well too bad. What exists on the main machine is what I have to deal with. I followed these instructions, but I seem to be unable to get things to work. In order to make the file server truly redundant, I cannot trust any one machine to hold the only data store on the network. Therefore, I have created a set of iSCSI drives on the VM-host of the main machine, and attached one to each file server VM. The end result is that I want my FS01 to sit on the heavy lifter, along with its iSCSI “drive”, and FS02 will sit on the secondary machine with its own iSCSI “drive” there as well. That is, neither iSCSI drive will end up sitting on the same machine as the other. As such, the clustered FS will utterly duplicate the contents of the iSCSI drives between each other, so that if one physical machine (or the FS VM) goes toes-up, the other has got a full copy of the data on its own iSCSI drive. My problem occurs when I try to apply the file server role within the failover cluster manager. Actually, it is even before that -- it occurs when adding the disks. Since I have added each disk preferentially to a specific VM (by limiting the initiator by DNS hostname, and by adding two-way CHAP authentication), this forces each VM to be in control of its own iSCSI disk. However, when I try to add the disks to the Disks section of Storage within Failover Cluster Manager, the entire process fails for a random disk of the pair. That is, one will get online, but the other will remain offline because it does not have the correct “owner node”. I mean, really -- WTF? Of course it doesn’t have the right owner node, both drives are showing the same node name!! I cannot seem to have one drive show up with one node name as owner, and the other drive show up with the other node name as owner. And because both drives are not “online”, I cannot create a pool to apply to a cluster role. Talk about getting stuck between a rock and a hard place! I’ve got more to add, but my work is closing for the day and I have to wrap things up. I will try to add more tomorrow morning when I get in. My main objective is to have a file server VM on each machine, the storage on each machine, but a transparent failover in case one physical machine fails. Essentially, a failover FS that doesn’t care which machine fails -- the storage contents are replicated equally on each machine. Am I even heading in the right direction?

    Read the article

  • Can't Delete Old Windows Directory

    - by David Mullin
    I got a new SSD drive for my computer, and have installed Windows on this drive. This left an old Windows directory on my old normal drive. I am now attempting to delete this old Windows directory, but am getting blocked by security. If I crawl down into each subdirectory, I can manually change the ownership and access rights for each file, but if I attempt to do it from the root directory, I get a "Failed to enumerate objects in the container. Access is denied" error. I have tried logging in as local Administrator, but this had the same effect. I figure that I am missing something stupid, but I just can't determine what it is.

    Read the article

  • How safe is it to run CHKDSK on an SSD?

    - by Eilon
    I recently saw Windows 7 pop up a warning or two that I should run chkdsk on my laptop. My laptop came with an SSD and I'm not sure if there are any negative implications to running chkdsk on such a drive. Are there any potential issues with reporting "bad sectors" on the drive? I would imagine that the physical concept of sectors is completely different between a platter and a microchip. I don't think my SSD supports TRIM. It's about 14 months old and a quick web search seems to hint that it doesn't (though it's nearly impossible to find out this info for sure!). I'm also not sure if TRIM is even relevant here since there shouldn't be much in the way of deletes. So, how safe is it to run chkdsk on my SSD drive? The model of SSD that I have is reported as "Samsung SSD PB22-JS3 2.5".

    Read the article

  • Copying files from NAS to NAS drives

    - by user1001421
    Very simple question. I've got 2 NAS drive that are "wire" connected via a router. If I have a wireless laptop and request a large amount of data be copied from one NAS drive to the other, does the network traffic go direct from the one drive to the other, going via the wired network, or does the network traffic go via my laptop, if you see what I mean. IE. From the NAS drives wired network, to the wireless network and then back to the wired network. Is this a common bottle-neck when copying a large amount of data? And if so, is there a way to avoid it. Thanks.

    Read the article

  • Disable Java Plugin in Google Chrome?

    - by Jeff Atwood
    This is the second time I've had a drive-by executable installed on my machine using the following: Google Chrome 6 (latest) Windows 7, UAC on This happened while I was browsing for images to add to a gaming.se post; one of the sites I visited (to get an image of a transfer cable) must have had drive-by browser exploit code running. UAC alerted me that a weird temporary executable wanted to run, and I declined, but I still got the fake antivirus executable running on my machine. Sigh.. I do have Java installed because I upload stuff monthly to clearbits.net and their uploader is a Java plugin. So my best guess is, websites are doing drive-by installs using the massive numbers of zero-day vulnerabilities in the Java browser plugins. For now, I have uninstalled Java, which works. But I wondered if I could disable the Java plugin in Google Chrome instead. So, how do you disable these vulnerable plugins in Google Chrome? I can't find the UI.

    Read the article

  • DFS share Access denied for some time after workstation is booted.

    - by Jonathan
    I have 2 Win server 2008 servers acting as domain controllers and Dns servers doing DFS. All my workstations are windows xp sp3. I have a DFS share drive mapped at login via login script. After the pc is booted I get an access denied message when trying to access the mapped drive. If i go to the file share on one of the servers i can get to it fine. After about 10 minutes i can try to access the mapped drive again and it works fine. I am not sure if it is some delay happening in dfs name resolution or a gpo issue.

    Read the article

  • How can I unfreeze an SSD connected to a remote server?

    - by chmac
    I don't have physical access to the machine, so I can't unplug the drive. # hdparm -I /dev/sda | grep frozen frozen The advice I've read elsewhere is to hotplug the drive, pull the power / sata cables while the machine is running. Those are not possible in this situation as I don't have physical access. I've tried power cycling the machine through the host's control panel a few times, but that hasn't worked. Is there any way I can unfreeze (unfrozen?) the drive without physical access?

    Read the article

  • Vertex 2 SSD is running faster than my Vertex 3 SSD?

    - by Kairan
    I used Acronis Disk Director to do a direct clone of my C:\ windows 7 x64 drive from my Vertex 2 to my new Vertex 3 SSD (Just to show the drive software winstall everything is identical.) I ran a performance test on Windows using the Windows Experience Index. The rating I am receiving when booting on the Vertex 2 is 7.5 While I am getting only a rating for the Vertex 3 of 6.9 My understanding is that the read/write speeds of the Vertex 2 is only up to 250MB/sec while the Vertex 3 is up to 500MB/sec. Copying a single file (3GB in size) from the Vertex 3 to itself was getting speed of approx 70-80MB/sec This speed is no better (maybe worse) than what I got from the Vertex 2 I am connected via the SATA 3 port on the motherboard, using an SATA 3 cable Is this issue caused by the drive cloning? Do I have a bad SSD?

    Read the article

  • How can I recover a huge folder that's been converted to a zero kb file on an NTFS partition?

    - by aalaap
    I have a 1TB drive with two 500GB partitions. One of them is being used as a Mac OS X Time Machine back up drive and the other one was NTFS and being used for storage. I had my entire 'iTunes Music' folder stored on it. Recently, there were some errors on the NTFS drive that caused chkdsk to run when in Windows, and it removed a lot of corrupt files. In this process, it converted my 'iTunes Music' folder into one zero KB file. How can I recover this? The partitions are intact and the other data on the disk is still accessible. It's just the 'iTunes Music' folder that's gone.

    Read the article

  • Device cannot be added on software-raid-1 array on Ubuntu 12.04

    - by George Pligor
    Unfortunately all tutorials I have found online until now on how to setup software-raid-1 are outdated on ubuntu 12.40 My target is to setup it on a system with a secondary disk drive that is already running. Format is not an option! I am trying to follow and adapt from 11.10 to 12.04 the following tutorial: http://www.howtoforge.com/how-to-set-up-software-raid1-on-a-running-lvm-system-incl-grub2-configuration-ubuntu-11.10-p2 On the above tutorial there is a successful command which creates a raid-1 array by setting the first disk drive with the installed system as missing: mdadm --create /dev/md0 --level=1 --raid-disks=2 missing /dev/sdb1 But when the time comes to add the first main drive with the installed system on the raid-array with this command: mdadm --add /dev/md0 /dev/sda1 I receive an error message. The error message says that the device /dev/sda is (which makes sense) busy! Note: hardware raid solution is not available since the system is a laptop with two disk drives! Thank you

    Read the article

< Previous Page | 275 276 277 278 279 280 281 282 283 284 285 286  | Next Page >