Search Results

Search found 68147 results on 2726 pages for 'context sensitive help'.

Page 816/2726 | < Previous Page | 812 813 814 815 816 817 818 819 820 821 822 823  | Next Page >

  • Blue screen while using windows even in safe mode

    - by FATE
    I have this problem with Windows 7 after I detected and cleaned up some viruses with Nod32. problem signature: Problem Event Name: BlueScreen OS Version: 6.1.7601.2.1.0.256.48 Locale ID: 1065 Additional information about the problem: BCCode: 50 BCP1: FFFFF900CA578010 BCP2: 0000000000000000 BCP3: FFFFF9600019619B BCP4: 0000000000000002 OS Version: 6_1_7601 Service Pack: 1_0 Product: 256_1 Files that help describe the problem: C:\Windows\Minidump\060312-34195-01.dmp C:\Users\Fatemeh\AppData\Local\Temp\WER-51215-0.sysdata.xml How can I prevent these crashes?

    Read the article

  • Executing batch file from sql server job

    - by uzay95
    I want to create backup job on sql server. And i want to execute batch file in job. I just wonder the part of executing batch file from sql job. Do you have any idea? Any help would appreciated. use MyDb go BACKUP DATABASE MyDb TO DISK = 'C:\BackUps\MyDb.bak' WITH differential go -- Call my batch file (which will zip MyDb.bak file)

    Read the article

  • VMWare Question

    - by Foghail
    Was wondering if anyone knew how to get 3-4 virtual box's to share the same monitor (i.e. so i could potentially get 4 box's on one screen, each screen environment would need to maintain its 1280x1024 internal resolution (so scale i guess?) anyways if you could help it would be greatly appreciated. I'm new to VMWare so please be gentle :-P

    Read the article

  • Executing batch file from sql server job

    - by uzay95
    I want to create backup job on sql server. And i want to execute batch file in job. I just wonder the part of executing batch file from sql job. Do you have any idea? Any help would appreciated. use MyDb go BACKUP DATABASE MyDb TO DISK = 'C:\BackUps\MyDb.bak' WITH differential go -- Call my batch file (which will zip MyDb.bak file)

    Read the article

  • Accessing Internal Server from Outside through SSH

    - by myahya
    The issue is as follows: I am using Ubuntu Linux. I am working from home, and I need to access a PostgreSQL database severer within my company's network. Call the DB server X. For security reasons, only one machine can be accessed from the outside using ssh, call it Y. So the scenario is: Home <--SSH-- Y <----- X(DB) I think something like SSH forwarding (configured in .ssh/config) could help with this, but I am not sure how to do this, any ideas?

    Read the article

  • use Amazon EC2 tools in Capistrano to get the servers to push the code

    - by APZ
    I am trying to use EC2 tools to get all the machines with a particular tag in some type of array in /config/deploy/prod.rb file in Capistrano. Something like this: In prod.rb file: //untested command workers-array[]=$(ec2-describe-instances -F vpc-id=1234 -F tag:Env=prod -F tag:SystemType=worker) for(i=0;i<workers-array.len;i++){ role :worker-A, workers-array[i] } I am not sure how we can do this in capistrano, am new to ruby too. Guys any help on this would be really appreciated.

    Read the article

  • Configuring an EH-WIC Card on Cisco 1941 Router

    - by Olanrewaju T
    I have a Cisco 1941 Router that has just two ports for Gigabit connection but wanted more so I got a four port Cisco EH-WIC Card and connected it to it. I have been trying to assign IP address to the port GigabitEtnernet 0/0/0 because I have a cable already connected to it whose device I want it facing the router directly because I dont want to NAT its address. I want to assign the physical address on the port. Kinldy help if you understand what I am saying. Regards

    Read the article

  • How do I filter certain javascripts from running and not block them all on Drudge Report?

    - by jay
    How do I block this particular "auto refresh" script on Drudge Report from running in my Firefox browser? I have NoScript and AdblockPlus plugins installed, but neither of them explain how to filter out a particular script and keep it from running and leave the rest alone. I don't want to stop all javascripts from running just the one listed below. Any help would be appreciated. "var timer = setInterval("autoRefresh()", 1000 * 60 * 3); function autoRefresh(){self.location.reload(true);}"

    Read the article

  • MongoDB and datasets that don't fit in RAM no matter how hard you shove

    - by sysadmin1138
    This is very system dependent, but chances are near certain we'll scale past some arbitrary cliff and get into Real Trouble. I'm curious what kind of rules-of-thumb exist for a good RAM to Disk-space ratio. We're planning our next round of systems, and need to make some choices regarding RAM, SSDs, and how much of each the new nodes will get. But now for some performance details! During normal workflow of a single project-run, MongoDB is hit with a very high percentage of writes (70-80%). Once the second stage of the processing pipeline hits, it's extremely high read as it needs to deduplicate records identified in the first half of processing. This is the workflow for which "keep your working set in RAM" is made for, and we're designing around that assumption. The entire dataset is continually hit with random queries from end-user derived sources; though the frequency is irregular, the size is usually pretty small (groups of 10 documents). Since this is user-facing, the replies need to be under the "bored-now" threshold of 3 seconds. This access pattern is much less likely to be in cache, so will be very likely to incur disk hits. A secondary processing workflow is high read of previous processing runs that may be days, weeks, or even months old, and is run infrequently but still needs to be zippy. Up to 100% of the documents in the previous processing run will be accessed. No amount of cache-warming can help with this, I suspect. Finished document sizes vary widely, but the median size is about 8K. The high-read portion of the normal project processing strongly suggests the use of Replicas to help distribute the Read traffic. I have read elsewhere that a 1:10 RAM-GB to HD-GB is a good rule-of-thumb for slow disks, As we are seriously considering using much faster SSDs, I'd like to know if there is a similar rule of thumb for fast disks. I know we're using Mongo in a way where cache-everything really isn't going to fly, which is why I'm looking at ways to engineer a system that can survive such usage. The entire dataset will likely be most of a TB within half a year and keep growing.

    Read the article

  • Changing the size of the Windows 7 taskbar

    - by dertoni
    Is there a way to change the size of the Windows 7 Taskbar? Internal or with the help of outside programs, both welcome. Something like the MacOS X Doc Zooming effect would be OK/nice, too. Edit: I'm essentially looking for a way to shrink it, because my laptop does not have a big screen so every pixel is valueable.

    Read the article

  • My server's been hacked EMERGENCY

    - by Grant unwin
    I'm on my way into work at 9.30 p.m. on a Sunday because our server has been compromised somehow and was resulting in a DOS attack on our provider. The servers access to the Internet has been shut down which means over 5-600 of our clients sites are now down. Now this could be an FTP hack, or some weakness in code somewhere. I'm not sure till I get there. How can I track this down quickly? We're in for a whole lot of litigation if I don't get the server back up ASAP. Any help is appreciated. UPDATE Thanks to everyone for your help. Luckily I WASN'T the only person responsible for this server, just the nearest. We managed to resolve this problem, although it may not apply to many others in a different situation. I'll detail what we did. We unplugged the server from the net. It was performing (attempting to perform) a Denial Of Service attack on another server in Indonesia, and the guilty party was also based there. We firstly tried to identify where on the server this was coming from, considering we have over 500 sites on the server, we expected to be moonlighting for some time. However, with SSH access still, we ran a command to find all files edited or created in the time the attacks started. Luckily, the offending file was created over the winter holidays which meant that not many other files were created on the server at that time. We were then able to identify the offending file which was inside the uploaded images folder within a ZenCart website. After a short cigarette break we concluded that, due to the files location, it must have been uploaded via a file upload facility that was inadequetly secured. After some googling, we found that there was a security vulnerability that allowed files to be uploaded, within the ZenCart admin panel, for a picture for a record company. (The section that it never really even used), posting this form just uploaded any file, it did not check the extension of the file, and didn't even check to see if the user was logged in. This meant that any files could be uploaded, including a PHP file for the attack. We secured the vulnerability with ZenCart on the infected site, and removed the offending files. The job was done, and I was home for 2 a.m. The Moral - Always apply security patches for ZenCart, or any other CMS system for that matter. As when security updates are released, the whole world is made aware of the vulnerability. - Always do backups, and backup your backups. - Employ or arrange for someone that will be there in times like these. To prevent anyone from relying on a panicy post on Server Fault. Happy servering!

    Read the article

  • [Disallow: /index.php] seems to block /my-beautiful-sef-url-123

    - by Jaroslav Záruba
    Hello I have robots.txt that looks like this: User-agent: * Disallow: /system/ Disallow: /admin/ Disallow: /index.php The obvious goal has been to prevent all the ugly URLs from being indexed, as they all begin with "/index.php". But for some reason all URLs like /my-beautiful-sef-url-123 are listed under Crawl errors in Google Webmaster Tools with "URL restricted by robots.txt". (When I test such URL it yields Allowed for both Googlebot and Googlebot-Mobile.) Can anyone help please?

    Read the article

  • Find what pages IE is fetching similar to console in firebug?

    - by Rob
    Firebug Lite bookmarklet is no help here, on half the pages it doesn't even show up. What I need is basically this: to show what IE is connecting to, because my application is working fine in firefox and chrome, but in an enterprise/corporate environment, I have to get it working with IE as well, and it's not working there. Any ideas of how I can get something similar just to see what the page is trying to (or not trying to) connect to? I'm using IE8 for this (on an windows XP machine, can't use IE9)

    Read the article

  • I am getting unmountable_boot_volume STOP 0x000000ED

    - by waterfalrain
    I read somewhere that the Ultimate Boot CD can diagnose computer problems. In this case would this program help me get my Windows XP back? When I turn on the screen and hit "Start Windows Normally" or go to Safe Mode I get a flash of the WinXP logo and then it goes to this error page. Can you tell from this message is it a hardware or software issue?

    Read the article

  • Avoid Windows Explorer to load complete executable file

    - by eli.work
    On Windows Vista, when browsing to a network folder containing executables, Windows Explorer seems to load all the files completely just to be able to show the executable icon (the resource monitor indicates loads of traffic during the loading of the directory) On XP only a part of the file is loaded. Is there a way to avoid the complete loading of these files? Note that disabling my anti virus does not help.

    Read the article

  • Looking for a way to diff many server's filesystem

    - by Itai Ganot
    There's a motivation in my office to make sure that all file systems on all servers in the production environment (It's a Windows environment) are identical until the last file and i'm looking for a program/tool which can help me achieve this goal. What i actually look for is a tool that will allow me to diff server's file systems by connecting them remotely (as they are spreaded around the world). Anyone knows a tool which allows it?

    Read the article

  • cURL Upload file AND send POST data

    - by kisplit
    Hello, I have a web server running some PHP that checks for an image (curl -F 'imageName=@myimage') and it also checks the POST data for username=&password=. When the PHP checks _REQUEST I can just do: curl -F 'imageName=@myimage' 'http://www.example.com/?upload=1&username=test&password=test' I need to instead check _POST for username and password due to specs. How can I upload the image and have the username=&password= post data? Any help appreciated!

    Read the article

< Previous Page | 812 813 814 815 816 817 818 819 820 821 822 823  | Next Page >