Search Results

Search found 13469 results on 539 pages for 'avoid trouble'.

Page 262/539 | < Previous Page | 258 259 260 261 262 263 264 265 266 267 268 269  | Next Page >

  • Norton Security Suite Symantec Download Manager Error: "Error writing to disk"

    - by Stephen Pace
    My broadband provider (Comcast) decided to switch their 'included with service' security suite from McAfee to Norton Security Suite. Their email directed me to a site that downloaded the Symantec Download Manager (NortonDL.exe) and that went fine. I'm running Windows 7 32-bit and running this application pops up the standard User Account Control message and the software is correctly identified as coming from Symantec. I answer 'yes' to allow the software to install and upon launch immediately get an "Error writing to disk" error. I searched the Internet for this error, but mainly I find Comcast users complaining about the same issue with no resolution other than to call Symantec. I found no one suggesting a successful workaround and it appeared that most of the support calls took up to three hours. I'd like to avoid that if possible. Ideas? To be honest, I'm getting close to bagging this installation and just moving to Microsoft Security Essentials.

    Read the article

  • How to safely remove a device blocked by the System process with a handle on \$Extend\$RmMetadata\$Txf

    - by Heinzi
    I have an external HDD which I would like to "safely remove". Unfortunately, my system (Windows 7 x64) complains that "the device is currently in use". Using Process Explorer I discovered which process is holding a handle on the device: Obviously, System is not a process that I can just kill and be done with it. I've done a bit of research and this seems to be a common problem, but no solution has been found so far (except for rebooting the machine, which I'd like to avoid if possible). Is there any solution to this problem that I've missed?

    Read the article

  • Navigating nested RDP sessions

    - by U62
    Often at work I have to RDP into one server in order to get to another. And no, there's no chance of the network getting sorted out such that I can connect to all machines directly :) I'm wondering if anyone has any good tips/tricks for working in this scenario. I tend to open RDP sessions full screen, so the way to get back to my desktop is to click the minimize or restore buttons on the toolbar the drops down from the top of the screen when you are using a full-screen RDP session. The problem is that when you log into A and then from there log into B, A's toolbar covers B's, so there's no easy way to get back to the desktop of A without logging out of B. At least as far as I can see. Maybe there's a keyboard shortcut that can pop you out of the innermost session? At the moment, I try to not use full screen mode on the nested session, but apart from having to remember to set that before I connect, it reduces the workspace and is something I'd like to avoid.

    Read the article

  • Can a virus corrupt a hard disk?

    - by sundar
    Shorter version: Hard disk corrupt, vendor claims warranty does not apply since it was "due to a virus" and "problems due to software are not covered under the warranty". Longer version: My Dell laptop recently refused to boot, and all attempts to 'repair' the Vista installation using the provided installation CD failed. I called up Dell support, and a representative took the laptop and after a day said the hard disk is corrupt. When I tried to ask for a replacement under the warranty, an official replied that the corruption was due to a virus, and "problems due to software are not covered under the warranty". Now, I get a doubt that he's trying to avoid having to provide it under the warranty. Is it possible for a hard disk to get corrupt due to a virus? If yes, is there any way we can detect it was due to a virus (as he claims to have detected)?

    Read the article

  • Moving SharePoint 2010 to a new domain

    - by Chris
    I have a small SharePoint 2010 farm (1 WFE / App Server & 1 SQL server) Our organisation is currently mirgrating to our holding company's global domain, so we now have a new local DC on site with trusts between the current domain and the new domain. I am going to need to move our SP Farm to the new domain and possibly rename servers to fit into the global naming convention (we are trying to avoid this at the moment, but might become a requirement) If there a way to script (stsadm / powershell) the user profiles and permission accross to the new domain? and on the server side, is it as simple as joining the servers to the new domain and updating all the service / farm accounts to accounts on new domain? I have googled this a bit, but everything I have found so far refers to MOSS 2007 or earlier. Any help / advise would be appreciated.

    Read the article

  • sharepoint administrative server not started causing sharepoint deployment instability

    - by Nathan
    When i deploy anything to a variety of sharepoint servers i get problems with some packages failing, from what i can see the only error given is as below. The timer job for this operation has been created, but it will fail because the administrative service for this server is not enabled. If the timer job is sched uled to run at a later time, you can run the jobs all at once using stsadm.exe - o execadmsvcjobs. To avoid this problem in the future, enable the Windows ShareP oint Services administrative service, or run your operation through the STSADM.e xe command line utility. Is this a known problem? I googled for the text but got only one other person with the problem who simply worked round it. I'm trying to batch deploy solutions programmatically and so these errors render the whole code worthless if you have to go back and redo bits by hand afterwards. Is batch deployment simply not possible? Thanks!

    Read the article

  • Using client certificates with wget

    - by Doc
    I cannot get wget to use the client certificates. The documentation speaks about using the --certificate flag. The use of the certificate flag is clear, I set it to use the PEM version of the client certificate. But when I connect I get the following error: HTTP request sent, awaiting response... Read error (error:14094410:SSL routines: SSL3_READ_BYTES:sslv3 alert handshake failure; error:140940E5:SSL routines:SSL3_ READ_BYTES:ssl handshake failure) in headers. Giving up. ssl handshake failure means the client did not supply a correct client cert. Still the client cert I use, works in a browser. Note: When I disable client authentication on the server, wget can connect. Note: The use of curl is suggested, but I'd like to avoid the switch.

    Read the article

  • How do I prevent IIS 8 from stopping idle ASP.NET applications?

    - by Lambo Jayapalan
    I have an asp.net application running on Windows 2012 in IIS 8 that has a very time consuming application start process (essentially the code running in the Application_Start() event can take up to 2 minutes). Thus I'd like to minimize the number of times the application is started so that the user can avoid a long wait. I've enabled Preload in the application settings, and I've set the Start Mode to AlwaysRunning in the application pool. Yet the application still ends after not being used for a while, resulting in a very long time for the first visit to the website after the application shuts down. Does anyone have any ideas on how I can prevent this? Thanks

    Read the article

  • Calculating Cloud Service Costs [closed]

    - by capdragon
    Possible Duplicate: Can you help me with my capacity planning? I would like to scale a web application to the Cloud and wanted to know if anyone had any experience calculating costs and could tell me how I would go about that. I have NO experience with cloud services at all. Currently my production environment consists of two web servers and one database server. If the application continues on a linear growth path, eventually, I want to scale to the cloud to avoid any more long term commitments to extra hardware. I want to be able to create a similar environment I have now as a baseline. Have this as my fixed cost that I will always have. I also want to calculate my variable costs that will increase with more users or bandwidth. I don't have a preferred cloud vendor. Amazon, Rackspace, Terremark or any other is fine as long as understand how to calculate my fixed and variable costs.

    Read the article

  • How can we achive a 403 Permissions Denied for a subdomain?

    - by marikamitsos
    We have a multisite installed in the root directory (multisite.com) and also a wordpress single installation on a subdomain (help.multisite.com) In the root .htaccess we placed: #START Security: Disallow access to folders Options All -Indexes # END Security On the main site (as expected we get) "403 Permission Denied. You do not have permission for this request /wp-content/blogs.dir/83/" Nice. :) BUT. We just noticed that when trying to access the subdomains folders we get: Internal Server Error. The server encountered an internal error or misconfiguration and was unable to complete your request.... Additionally, a 500 Internal Server Error error... This is something we do NOT want. So the question is: How can we avoid the above result and make the message for the subdomains be "403 Permission Denied" (the same as for the main site and NOT "500 Internal Server Error" (as it is now)? We put what, where?

    Read the article

  • VMWare Player guest does not re-aquire IP on bridged interface after host loses and re-acquires network connectivity

    - by Vineet
    I am running a Ubuntu Linux image in VMWare Player on my laptop. The host is Windows XP. There are two network adapters configured in VMWare Player - one is host-only and the other is bridged. Everything works fine as long as the host does not lose network connectivity. If the host loses connectivity (even briefly) and then re-acquires it, my bridged adapter in VMWare Player is still unable to get an IP address. The host-only adapter remains unaffected. Disconnecting/Reconnecting the bridged adapter does not help. Simply restarting the guest OS does not help. Restarting VMWare Player does not help. The only remedy is to reboot the host and then bring up VMWare Player all over again. Is there something I can try to avoid this reboot? I searched the existing questions but they seemed to talk about adapters in NAT mode, whereas I am interested in retaining Bridged mode.

    Read the article

  • In Nginx can I set Keep-Alive dynamically depending on ssl connection?

    - by ck_
    I would like to avoid having to repeat all the virtualhost server {} blocks in nginx just to have custom ssl settings that vary slightly from plain http requests. Most ssl directives can be placed right in the main block, except one hurdle I cannot find a workaround for: different keep-alive for https vs http Is there any way I can use $scheme to dynamically change the keepalive_timeout ? I've even considered that I can use more_set_input_headers -r 'Keep-Alive: timeout=60'; to conditionally replace the keep-alive timeout only if it already exists, but the problem is $scheme cannot be used in location ie. this is invalid location ^https {}

    Read the article

  • Skipping hardlinks when using TSM Backup

    - by Lars Haugseth
    We need to backup a filesystem with lots of hardlinks. Since there are several hardlinks for each "true" file, we would like to skip all the hardlinks when backing up the filesystem to avoid n exact copies of each file. The backup is done using Tivoli Storage Manager Backup, and we've been unable to get it to treat hardlinks as anything other than separate files to be backed up alongside each other. In case it's relevant for possible solutions, I'd like to note that it's possible to tell a hardlink from a proper file by the filename: foobarbaz-123.ext # file foobarbaz-123-1.ext # hardlink foobarbaz-123-2.ext # hardlink barbazfoo-456.ext # file barbazfoo-456-1.ext # hardlink barbazfoo-456-2.ext # hardlink barbazfoo-456-3.ext # hardlink That is, all hardlinks have two hyphens in the filename, where as proper files have just the one. The server is running Ubuntu Linux, and the files are situated on a gfs volume on our SAN.

    Read the article

  • My Intel Core i3 frequently shuts down due to "Processor Thermal Trip error"

    - by explorest
    My Desktop PC running Windows 7 on Intel Core i3 dual processor machine started shutting down during the boot process a couple of days ago. I left it alone for about 48 hours and tried again. The boot process informed me about Processor Thermal Trip as being the cause of boot failures and suggested a Windows recovery process. I went through the recovery process which completed without problems. The PC booted normally and Windows 7 worked as expected but after 10 minutes it again tripped and now whenever I start it, it shuts down during the boot process. What should I do next?... is there a boot menu / setup settings that needs to be adjusted to avoid it... Or is there a hardware problem (which seems unlikely as the machine occasionally boots perfectly )?

    Read the article

  • Raspberry Pi: mvoe home & var folders to USB HDD?

    - by doni49
    I have a Raspberry Pi running raspian. I've installed Apache2, PHP & MySQL. Apache & MySQL are both configured to use sub-directories of /var. I'd like to use a directory on my USB HDD instead of my SD card. I'd also like to move the /home directory to the USB HDD. I'd like to avoid re-partioning my HDD if possible. I thought maybe I could use a symlink to tell raspian that /var is really at /media/USBHDD1/var and /home is really at /media/USBHDD1/home. I tried it last night but couldn't get it to work.

    Read the article

  • Changing Network Path of Offline Files

    - by Adam
    Many of our users have their Home folder set as Available Offline. Their Windows 7 laptops will not be back on our network for a few weeks. In the mean time, we're setting up new servers and reorganizing our files, so the network path to the Home folder is going to be completely different. Based on some testing I did, when the users return, any files they've created or modified while offline will be gone, and the new Home folder will be there and not set to sync. The offline cache of the old Home folder is still accessible through the Sync Center, but they're not going to want to dig through that and try to find what's missing. Avoiding this would involve keeping the old server around and moving everyone to the new location in person, so we know for sure they're synced first. Is there any way to avoid this that isn't as tedious, like a quick registry edit or something that will point the old offline cache to the new location?

    Read the article

  • X-FI optical output not being recognized.

    - by Luciano
    Hello there! I have an issue and I don't know what to do! I have a SoundBlaster X-FI Xtreme Gamer, which has a spdif mini-optical output. I have it connected to a home theater, via a toslink cable and an adapter to fit in the mini plug. It used to work perfectly, but now it doesn't anymore. There is light travelling through the cable, but the HT doesn't pick up the signal. The sound card of the motherboard also has an optical output (regular, not mini) and it works ok, so the cable and the HT are fine, I think. What else should I try? Thanks, Luciano. PS: I only enabled the onboard card for testing, but usually it's disabled to avoid any kind of conflicts. PPS: I'm using Windows 7 64bit with latest drivers of everything, but like I said. It used to work.

    Read the article

  • Is there a free-embedded SSH solution ?

    - by ereOn
    Hi, I'm working for an important company which has some severe network policies. I'd like to connect from my work, to my home linux server (mainly because it allows me to monitor my home-automated installation, but that's off-topic) but of course, any ssh connection (tcp port 22) to an external site is blocked. While I understand why this is done (to avoid ssh tunnels I guess), I really need to have some access to my box. (Well, "need" might be exagerated, but that would be nice ;) Do you know any web-based solution that I could install on my home linux server that would give me some pseudo-terminal (served using https) embedded in a web page ? I'm not necessarily looking for something graphical: a simple web-embedded ssh console would do the trick. Or do you guys see any other solution that wouldn't compromise network security ? Thank you very much for your solutions/advices.

    Read the article

  • AD Custom Attribute with unique value

    - by Zilog
    I have custom AD attribute added to my AD schema. Attribute's syntax is Unicode String. It is added to user class and the purpose of it is to store user ID of corporate ERP system. Problem is, that I can have two or more user objects with the same value stored in that attribute, which is something that I would like to avoid. Is there a way to configure AD attribute so it is unique within the domain boundary? (The same behavior as SAMAccountName attribute.) i.e : If there already exists user object in AD with this attribute set to "JSmith" and I try to set that attribute with the same value for another user, Direcory Services will refuse to update that object and give me "already exists" error. Domain functional level is windows 2003.

    Read the article

  • Proper way for changing MAC address in a linux VM?

    - by HappyDeveloper
    I tried to change the MAC address in a ubuntu VM (virtualbox), but after that it threw lots of errors during boot, and then I had no internet connection. Then I saw that the interface was renamed to eth1, so I edited /etc/network/interfaces to change eth0 to eth1, rebooted (didn't know how to restart the network), and boot was now faster and internet worked fine. But now after every time I log in, I get 1 or 2 error messages that say nothing, they only ask me if I want to report them. So I was wondering, is there was a proper way to change MAC address, to avoid these issues?

    Read the article

  • Switch computer off automatically when the internet is disconnected

    - by Vin
    Is there any free software that will detect the status of internet and shutdown the system when the internet disconnects? Or (more generally) can something run a task when the internet disconnects, and I can route it to shutdown the machine? I am downloading things overnight. But my problem is that sometimes the internet connection will be disconnected in the middle of the night, so there is no point in keeping the system running the rest of the night. So I need to avoid this problem by detecting the net connectivity.

    Read the article

  • bind9 dns proxy

    - by Zulakis
    We are offering multiple SSL-enabled services in our local network. To avoid certificate-warnings we bought certificates for server.ourdomain.tld and firewall.ourdomain.tld. We now created a zone in our local DNS-server in which we pointed the hosts to the corresponding private-ips. Now, each time another record for ourdomain.tld, like for example www.ourdomain.tld or alike are changed, we need to update it on both our public-dns-server AND the local dns-server. I would like our local bind-dns to serve all the information from our public-dns but serve different information for these 2 hosts. I know I could possibly have our private-ips in our public-dns but I don't want that for security reasons. The internet dns-server is being managed by a third party, while we have full control of the intranet one. Because of this I am looking for a solution which lets the intranet retrieve the records from the internet one.

    Read the article

  • What should I know before I set up RAID 6 on Linux?

    - by Dan Ellis
    I just ordered five 1TB drives to install as a RAID 6 array in a Linux server (keeping the existing 1TB drive as a boot disk). I want to use Linux MD for RAID rather than a RAID card, to avoid lock-in. The intended use is for storing filesystems for Xen development environments and an AFP server for iPhoto/Aperture/Lightroom. What things should I know before I set it up? For example, what would be a good choice of filesystem, and what chunk size should I use?

    Read the article

  • Moving VMWare Fusion image to Boot Camp

    - by Kristopher Johnson
    I have Windows 7 64-bit running in VMWare Fusion on my MacBook, but am disappointed with the performance, and so I want to try Boot Camp. However, I'd like to avoid reinstalling Windows and all my applications; I just want to somehow copy my VMWare Fusion "disk image" to a Boot Camp partition. My initial thoughts are that I should be able to run a Windows backup program in VMWare Fusion to back up the entire virtual disk, then set up Boot Camp and restore from that backup. However, Googling finds a few posts by people who have tried that and have encountered problems. So, is there a "known good" procedure for doing this?

    Read the article

  • Command line safety tricks

    - by deadprogrammer
    Command line and scripting is dangerous. Make a little typo with rm -rf and you are in a world of hurt. Confuse prod with stage in the name of the database while running an import script and you are boned (if they are on the same server, which is not good, but happens). Same for noticing too late that the server name where you sshed is not what you thought it was after funning some commands. You have to respect the Hole Hawg. I have a few little rituals before running risky commands - like doing a triple take check of the server I'm on. Here's an interesting article on rm safety. What little rituals, tools and tricks keeps you safe on the command line? And I mean objective things, like "first run ls foo*, look at the output of that and then substitute ls with rm -rf to avoid running rm -rf foo * or something like that", not "make sure you know what the command will do".

    Read the article

< Previous Page | 258 259 260 261 262 263 264 265 266 267 268 269  | Next Page >