Search Results

Search found 59959 results on 2399 pages for 'time complexity'.

Page 506/2399 | < Previous Page | 502 503 504 505 506 507 508 509 510 511 512 513  | Next Page >

  • Outlook.com Sweep rule vs Filter rule

    - by Mahesha999
    New outlook.com introduced new sweep option. With it, you can create a filter to move messages from particular sender to a particular folder after certain days the mail comes. This rules gets added to the Filter list. However we cannot edit this rule by clicking Edit in front of it in that list. The dialog appears saying you should delete the rule and create the new one. I don't understand why sweep rules are considered differently from Filter rules? I have 68 filter rules. What if I want to add such time clause to these filter rules so that instead of moving the incoming mails immediately to a certain folder, they will be kept in Inbox for say 10 days and then they will be moved. Such time clause cannot be added to the filter rules. So should I delete all these 68 filters and create new Sweep rules?

    Read the article

  • What is the right way to use Internet Sharing in OS X?

    - by Bob King
    For a while now I've been using the Wireless in my Mac Mini as a seperate WiFi access point because my main router doesn't quite cover my whole house. There is a dedicated CAT 5E line down to the Mini, which has a static IP address. I've turned Internet Sharing on, using Airport/WiFi but it seems my options for security are limited. Can I do WPA2? Also, it seems like my iPhone only connects 25% of the time, and the rest of the time it claims I've given a bad password. I've tried "Forget this network" repeatedly.

    Read the article

  • Having to Toggle Between Two Sets of Video Drivers

    - by Maarx
    So, I've got a Win7 64-bit gaming PC with GTX 260's. Recently, StarCraft 2 had an issue with flickering, which NVidia fixed with a new set of drivers. However, these new drivers induce unplayable graphical errors with Neverwinter Nights 2, something me and my friends still play from time to time. I am seeking advice on the "best" way to rectify this situation, to be able to switch between two driver releases, without compromising the stability of my system (if Windows stability isn't an oxymoron). I'm wondering if Windows 7 is structured in such a way that I can constantly reinstall these two sets of drivers back and forth overtop each other, possibly six or eight times a day, without very quickly driving myself to reformat to maintain that "just like new" performance. I'm loathe to have to reformat the drive and maintain two copies of the operating system, but I'll do it if I have to.

    Read the article

  • change browser to open weblinks in thunderbird

    - by Bevor
    Hello, because it's almost obvious that either Thunderbird or Firefox freeze my whole system after some time, I'd like to not use FF at the moment and let Thunderbird run only for a short time as long as I check e-mails. Thunderbird web links should be opened with opera from now. For that I went in Thunderbird to Preferences-Advanced-General-Config Editor and set the following: network.protocol-handler.app.http -> /usr/bin/opera network.protocol-handler.app.https -> /usr/bin/opera network.protocol-handler.warn-external.ftp -> true network.protocol-handler.warn-external.http -> true network.protocol-handler.warn-external.https -> true Moreover I changed in Gnome in System-Preferences-Prefered Applications Opera as webbrowser and I checked in FF that it should warn if FF is not the default browser (obviously it isn't anymore). Unfortunately all of these settings doesn't work. Thunderbird still opens web links with FF. Any explantions why? (I already restarted Thunderbird although I don't have to, but no effect)

    Read the article

  • Trick to enabling FTP upload from an app in Windows 8?

    - by Craig
    I have an older Win32 app that uses FTP to upload some data to a server from time to time. This app works great in Windows XP and has been working fine in Windows 7 and Vista. For some reason, when running under Windows 8 it has trouble uploading its data. It appears to connect and upload the data OK but it has times out when disconnecting. The data does not appear on the server. I have limited visibility into the app while it's running so sorry for the vague nature of the description. My question is: Is there some security setting in Windows 8 that would keep this app from using FTP? When I check the settings it looks to me like I have it set up to prompt me about any permissions problems it might have, and I'm not seeing any prompts. I've tried running this app in XP/SP3 compatibility mode with no luck. At this point I'm just looking for some clues as to where to look.

    Read the article

  • Strange sound from Laptop, Microphone, Speakers or Soundcard?

    - by David Schilling
    I have recently purchased a second hand computer. There didn't seem to be anything wrong with it on purchasing, however recently I have noticed a small issue. Any time I play sound or listen through my headphones, I can hear what seems to be the CPU working (ie when beginning a new task, there is a small crackling). Also if I tap my laptop I can hear the sound louder, as if it is being picked up by the mic and then playing that sound in real time through the speakers or headphones. I have tried reading other topics but cannot find an answer to this question. These are the specs I am running on; Windows 7 Home Premium 32bit Pentium Dual Core CPU T4300 @ 2.10Ghz 2.10Ghz I hope someone can help me with this issue, it would be much appreciated.

    Read the article

  • How to configure DD-WRT routing table when creating an isolated network segment for PCI C VT compliance

    - by tetranz
    I'm the volunteer support and system admin person at a small private school. We need to setup a PCI compliant Windows PC as a virtual terminal for credit card processing. I've read questionnaire SAQ C-VT and, to quote, this computer needs to be accessed: "via a computer that is isolated in a single location, and is not connected to other locations or systems within your environment (this can be achieved via a firewall or network segmentation to isolate the computer from other systems)" Our setup is as follows: DSL modem from ISP is setup to be a "transparent pipe" with no extra services. That goes into the WAN port of Linksys WRT54-GL running a DD-WRT. The LAN is 192.168.1.x. There are a couple of other WRT54-GL / DD-WRT devices. One is used as a wireless AP and another is a client bridge. To isolate the VT (virtual terminal) machine, I have another DD-WRT device. Its WAN is connected to a port on the 192.168.1.x LAN. The virtual terminal machine is connected to its LAN which is at 192.168.10.x. The SPI Firewall etc is turned on. It's basically the default DD-WRT gateway setup where the "ISP" is our own LAN. That's working. All incoming traffic to the VT machine is blocked, including from our own LAN. The VT can access the internet BUT, and here's the problem, it can also ping any of the computers on the 192.168.1.x LAN. I think I need to stop that. I'm guessing that I could do something with the Static Routing table in the VT machine's DD-WRT device. I need to route anything going to 192.168.1.x other than the gateway which is 192.168.1.1 to 0.0.0.0 or something like that. That's where I'm stuck at the end of my knowledge. Or ... do I need to get yet another DD-WRT so the network is "balanced". Maybe I need to have the internet from the DSL going into a DD-WRT which has only two devices on its LAN i.e., two other DD-WRTs, one for the main LAN and one for the VT. I think that would do but I'd like to avoid the extra cost and complexity if I don't need it. Thanks

    Read the article

  • Resume ./configure after solving dependency requirement

    - by pagliuca
    Sorry about my bad English. Here's the question: Is there a way to resume ./configure script from where it stopped just after fixing a dependency requirement? It seems that the script runs for a long time just until it throws me another unsatisfied dependency... After I fix it, the ./configure script starts running from zero again. This would sure save me a few minutes every time I compile a package which I don't have easy access to its dependencies. Does this resuming depend on the publisher providing caching for its configure script, or is there something I can do about that? Any ideas? Thank you.

    Read the article

  • How to get notifications from Windows Update when optional updates are available?

    - by Tom
    On my Windows 7 I set windows update to "download updates, but let me choose whether to install them". The problem is I don't get a notification when an optional update is available like definition updates for microsoft security essentials. I have "give me recommended updates the same way I receive important updates" checked, but I have no notification balloon or icon when there is a new optional update is available. From time to time I open Windows update and then I see I have new optional updates available. How to get notifications for those too?

    Read the article

  • Linux: don't use file system cache under a directory

    - by GetFree
    For a PHP website I'm monitoring, I need to see what files are being used each time the browser makes a request. I thought of using find . -type f -amin 1. With that I get all files which were read in the last minute (it's a developing server so only I am using the website). I took care of removing the noatime attribute from the mounting point. However there must be something else that's preventing the kernel from reading the actual files on disk because the access time is not being updated when I read a file. I guess it must be the file-system cache which is retrieving the files from memory. Is there a way to disable file caching under a specific directory? (public_html in my case) Also I read somewhere that there is the nobh mounting atributes which apparently disables file caching under that mounting point, but I'm not sure.

    Read the article

  • Monitor does not work past certain resolution.

    - by Steve Stifler
    I have a Macbook 5.1 with Snow Leopard, and a Samsung SyncMaster 2253BW that I use for an external monitor. I connect it to my laptop via a DVI-to-MiniDisplayPort adapter. Whenever I try to set the resolution of the monitor to 1680x1050, its maximum resolution (at least the only one that fills the entire screen), the screen flickers once or twice before claiming there is no source attached. I've had this issue since the first time I tried to connect the two, however for some reason for a brief period in time the monitor would successfully display at full resolution. All lower resolutions work fine too, it's just at 1680x1050 that the problems occur. Any ideas as to why this is happening? Thanks.

    Read the article

  • Staggering java linux process startup to prevent OOM

    - by ctennis
    I am running a number of java processes on a single Linux machine. From a memory and computing standpoint, everything is fine when things are static. However, periodically we use a configuration management package up upgrade the jar or war files, and restart the java process. The problem is, that is restarts them all relatively quickly, and so we get 10 or so java VMs restarting all at the same time (we use daemontools for the service stops/starts), which wreaks havoc on the machine, in terms of OOMs or just really slow. This is because it's spawning the JVM 10x at the same time. Other than trying to stagger the startups, is there a smarter way of handling this? Maybe a sysctl tuning performance parameters, or a JVM parameter?

    Read the article

  • Is it possible to upload only files that have been updated into a server?

    - by kamikaze_pilot
    Hi guys, Suppose I have a server accessible via FTP and it hosts websites Suppose I want to edit the website locally so it wont affect the site live, and suppose I edit a whole bunch of files, and I don't want to deal with the hassle of keeping track of which files I've edited all the time... Once I finished editing I want to upload it to the server via FTP....is there some FTP software that automatically detects which files have been edited and have only those files uploaded and overwritten rather than having me manually choosing the files I've edited (and hence having to keep track of edited files) or have me upload the entire site which is a waste of time thanks in advance

    Read the article

  • System failure - need diagnostic recommendation

    - by Ladislav Mrnka
    I have big problem with my computer. Configuration is: Intel i7 + 6x2GB OCZ DDR3 Motheboard: Asus P6T Deluxe V2, HDD controller configured to AHCI Main drive: OCZ Vertex 2 (SSD) - contains all installed programs and system Second drive: Samsung SpinPoint - contains User profiles, ProgramData, virtual machines and databases Third drive: Samsung SpinPoint - data drive + backups OS: Windows 7 Ultimate x64 I have never had any problem with this computer until now. During weekend my computer completely crashed without any reason. Each time I tried to boot to Windows I got BSOD with message BAD_SYSTEM_CONFIG_INFO and automatic restart (I didn't install any new SW or HW). But after restart main OCZ drive was disconnected (not detected by BIOS). When I turned off and on computer, the drive was again connected. It also happend every single time I tried to repair installation somehow. It ended with some error and after restart drive was disconnected. The only thing which worked was format + fresh install. After installing almost everything I wanted to install Visual Studio 2010 Ultimate (complete installation without SQL Server Express). During installation of VS itself I always get BSOD - it is too fast so I'm not able to read description. After restart it searches for all disk drives for really long time and sometimes it changes boot drive so the system is not able to start - Bootmgr not found. After reconfiguring BIOS the system starts. There is no event describing the failure in Event viewer. Installing VS 2010 is absolutely necessary for me. I need help with diagnostic. I need to find where is the problem - I expect that the problem is in OCZ drive or in HDD controller on motherboard but I don't know how to find it. All components still have valid warranty. Can you recommend me some approach or tools to find the problem? Edit: I'm still looking for source of the problem. New information is that Windows are not able to perform check disk (Chkdsk) on the SSD system drive. After restarting it always starts checking drive and in part where files are checked it fails with BSOD - BAD_SYSTEM_CONFIG_INFO. After next restart and skipping check disk tests the system runs.

    Read the article

  • Should I use UPS with my laptop to extend the battery life?

    - by Mehper C. Palavuzlar
    I recently bought a new Vaio laptop which has a 4400 mAh battery. I will use it at home most of the time, and I have an unused uninterruptible power supply (UPS). Should I remove the battery and connect my laptop to UPS? If I do that, how much does it effect the battery's life? I know that batteries stand longer when kept in half-charged in a cool place, but their lifespan decreases in time due to age effect as well. So is it worth to keep the battery out of my laptop, use a UPS instead and mount it when necessary? or should I continue to work with battery mounted?

    Read the article

  • How can I run a job when the server load is low?

    - by jberryman
    I have a command that runs a disk snapshot (on EC2, freezing an XFS disk and running an EBS snapshot command), which is set to run on a regular schedule as a cron job. Ideally I would like to be able to have the command delayed for a period of time if the disk is being used heavily at the moment the task is scheduled to run. I'm afraid that using nice/ionice might not have the proper effect, as I would like the script to run with high priority while it is running (i.e. wait for a good time, then finish fast). Thanks.

    Read the article

  • let mysql use full resources

    - by shekhar
    We have a Windows server 2008 R2. It has 24GB RAM and 2.926 Mhz, 8 Core(s), 8 Logical Process. We have a large MySQL database and it's taking a lot of time to execute some queries. But while running queries, I have observed it's NOT utilizing full resources. I am thinking that if I let MySQL utilize maximum resources on my server, it can reduce execution time. How can I let my MySQL server utilize maximum resources in a healthy way?

    Read the article

  • Decreasing lagging on router, while gaming

    - by user2699451
    I had absolutely no idea where to post this question and get a professional answer for it but here goes... Okay, so I guess everyone whos is reading this had played online, and so I was playing LoL again tonight and my brother decided that now was a great time to go on youtube and start watching a movie, so my ping (connecting from South Africa to EU west server) is around 190-220 average, however it started spiking to 2000 and average was 600-800, so it arised the question, how ther hell can I "kick" him off for the time being I tried reasoning it out with him but its like playing chess with a pigeon, he's studying to be an engineer, and I just cant win an argument with him, so i need to step it up a level... I have in the past used the aireplay method by sending deauth packets but it only helped so much, is there another way of either kicking a peer of the local wifi or decreasing the lag spikes while in session or even splitting the bandwidth equally in 2 or 3,etc What do I do p.s. sorry if off topic, if it is not appropriate, just say which website will be able to help or assist me...

    Read the article

  • Why does Windows share permissions change file permissions?

    - by Andrew Rump
    When you create a (file system) share (on windows 2008R2) with access for specific users does it changes the access rights to the files to match the access rights to the share? We just killed our intranet web site when sharing the INetPub folder (to a few specified users). It removed the file access rights for authenticated users, i.e., the user could not log in using single signon (using IE & AD)! Could someone please tell me why it behaves like this? We now have to reapply the access right every time we change the users in the share killing the site in the process every time!

    Read the article

  • System will only boot into a Live CD with core disabled in BIOS and ACPI off.

    - by CookieOfFortune
    I have a system that was running a Q6600 and Windows 7 RC. It crashed last night into a BSOD with a MACHINE_CHECK_EXCEPTION. Now it does that every time it boots into Windows. I tried using my Ubuntu Live CD, but the kernel would error out, the stack trace showed something along the linse of "Can not synchronized to one of the CPUs". Working from this, I enabled the "Limit CPU to 3 cores" option in my BIOS and tried again. This time, it seemed to have died after an ACPI call, so I disabled that during the boot and now it is running from the Ubuntu Live CD, showing 2 cores. Does anyone here think I have any hope or is it simply a CPU waiting to die? EDIT: 1 core now.

    Read the article

  • bash shell script which adds output of commands

    - by John Kube
    Let's say I have a command called foo which prints a number to the screen when called: $foo 3 Let's also say I have another command called bar which prints another number to the screen when called: $bar 5 I'm looking to write a shell script which will add together the output of foo and bar. How would I do that? (The outputs from the commands are not known ahead of time. They just so happen to have been 3 and 5 the last time they were run. They could have been something else.) Thanks!

    Read the article

  • tail -f updates slowly

    - by Cliff
    I'm not sure why, but on my Macbook Pro running lion I get slow updates when I issue "tail -f" on a log file that is being written to. I used to use this command all the time at my last company but that was typically on Linux machines. The only thing I can think of that would possibly slow the updates are buffering of output and/or maybe a different update interval on a Mac vs. Linux. I've tried with several commands all which write to stout relatively quickly but give slow updates to the tail command. Any ideas? Update I am merely running a python script with a bunch of prints in it and redirecting to a file vi " my output.log". I expect to see updates near real time but that doesn't seem to be the case.

    Read the article

  • Hyper-V Virtual Disk Creation Taking Forver

    - by mnemosyn
    After some struggle, I finally managed to set up Hyper-V 2008 R2 on our server. So I connected to it using the Hyper-V Manager from a Windows 7 client and used the "New Virtual Machine Wizard". I set up a 350GB virtual hard disk. So I hit the "finish" button and the Hyper-V manager has been working for 24hours now, showing merely a dialog "Creating Disk". A console on the Hyper-V still reports 99.9% free space on the HD, but the machines HD LED flashes from time to time (making a rather idle impression, it's not flashing frenetically). Does this usually take this long? Is there a way to find out whether it's still working or just idling? Should I repeat the process? Guides on the net tell me to be patient, but 1d seems a bit extreme!?

    Read the article

  • WIth more mobile users, my geo ip database is becoming useless.

    - by Marius
    Hello there, I've been enjoying the benefits of Geo IP lookup from database for some time. Its great. People are increasingly trying to access my site from a mobile phones or 3G modems, and their physical location seems to have little relation to whereabouts my IP lookup tells me they are. A user who is on the east cost of my country, may be looked up as being in the far inland, or up north. And one user may be reported as being in one location in one moment, and seconds later, be 100s of kilometers away. This is becoming a problem, and I need to find a solution. I am already updating my database monthly, but it has little effect. What can be done? Thank you for your time. Kind regardsMarius

    Read the article

  • teamviewer display is slow to refresh

    - by brunoais
    I'm using teamviewer from time to time to connect my portable computer with my desktop computer when I need to check on something (personal use). One of the main things I've been noticing is that it is quite slow transfering the screen updates that keep happening. This happens more frequently when I'm moving full windows or when I open a window or when watching a small movie clip. I've tested the lowest maximum p2p speed I ever got in the last week between both and it is the desktop's upload speed of 6MB/s with the ping of 40ms (usually ~15 ms). With this I was expecting to have very sharp image and fast transfer speeds which is not what is happening. Just to clarify, both computer's individual core's CPU is not reaching its peak (25%) while teamviewer is active, so the encryption and decryption is not the cause. Why my connection between both computers so slow?

    Read the article

< Previous Page | 502 503 504 505 506 507 508 509 510 511 512 513  | Next Page >