Search Results

Search found 23890 results on 956 pages for 'issue'.

Page 287/956 | < Previous Page | 283 284 285 286 287 288 289 290 291 292 293 294  | Next Page >

  • How to re-run one line of a PowerShell cmdlet if it fails

    - by pansal
    I wrote a PowerShell script which is supposed to send emails automatically, but sometimes the email won't send out due to a network issue. Here is how I'm sending email: $smtp_notification.Send($mail_notification) Here are the error logs: Exception calling "Send" with "1" argument(s): "Failure sending mail." At line:1 char:24 + $smtp_notification.Send <<<< ($mail_notification) + CategoryInfo : NotSpecified: (:) [], MethodInvocationException + FullyQualifiedErrorId : DotNetMethodException Is there anyway to re-run the sending line when I met this failure? Can anyone give me some suggestions please?

    Read the article

  • Terminal services and screen savers

    - by Corsair
    I have a Window Server 2008 R2 server acting as a terminal server for mobile devices running a single application, but I am having issues with the screen saver popping on after ten minutes. Is there a setting in terminal services to make the screensaver settings universal? If not, how should I approach this issue?

    Read the article

  • Flash player 11.4 hungs firefox immediatly after start

    - by killer_PL
    I have installed Flash player 11.4.402.287 on a "fresh" PC, moments after installing Windows, drivers, directX etc. After installing Flash Player Firefox totally hungs immediatly after run. First page (adobe page, because it was launched before closing firefox) doen't load. I see new tab but FF goes hung to zero-responsivity, even loading icon stops in single position. KMPlayer, wich uses firefox also hungs. How to solve this issue ?

    Read the article

  • What might cause the Save and Open dialog windows to just not work?

    - by Nick Bolton
    In several applications, the Save and Load dialog windows are just not showing. And in notepad (which obviously handles the return code) is reporting an out of memory issue, which I'm sure is not the case; I think it's assuming it's out of memory as it can't get the window handle. In any case, there's something definitely wrong with Windows, but there's nothing in the event log. Any idea why this might happen?

    Read the article

  • Consumer Priority Service - How many years of coverage should I buy for my Apple MacBook Pro 17"?

    - by jaffsmith
    I recently purchased a Apple MacBook Pro 17" online. It costs me $2299 plus shipping. If the unit should have an issue over the next few years I want to be fully covered so I will be purchasing the Consumer Priority Service extended warranty as well. The online dealer offers a 2 year or a 3 year for about $100 more, Is it worth the extra money for the 3 year warranty?

    Read the article

  • Google Chrome and Firefox Rendering

    - by user13503
    When attempting to visit sites running javascript, Google Chrome and Firefox often fail to render them properly. For example, when hitting google.com/codesearch results, I just get a blank page. http://img340.imageshack.us/img340/8527/200910071518.png This rendering problem happens in both browsers on a machine running Windows XP Professional x64. Internet explorer is able to render the page just fine. Also, controls on other pages sometimes fail to respond, for example in Google Docs. Has anyone encountered this issue before in Chrome/Firefox? Do they share a common installation of the V8 javascript engine? I'm baffled by this issue. Here's the source code for the image above that Chrome loads when attempting to retrieve the page: <!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN" "http://www.w3.org/TR/html4/loose.dtd"><html><head><script>function a(c){this.t={};this.tick=function(d,e,b){var f=b?b:(new Date).getTime();this.t[d]=[f,e]};this.tick("start",null,c)}var g=new a;window.jstiming={Timer:a,load:g};try{window.jstiming.pt=window.external.pageT}catch(h){}; </script><title></title><script language="javascript" src="/codesearch/js/CachedFile/F1A2CB189D0FCB1FF201C42BF6A5447C.cache.js"></script></head><body><iframe src="javascript:''" id="__gwt_historyFrame" style="width:0;height:0;border:0"></iframe><script>if(window.jstiming){window.jstiming.a={};window.jstiming.c=1;function k(a,d,f){var b=a.t[d];if(!b)return undefined;b=a.t[d][0];if(f!=undefined)var h=f;else h=a.t.start[0];return b-h}window.jstiming.report=function(a,d,f){var b="";if(window.jstiming.pt){b+="&srt="+window.jstiming.pt;delete window.jstiming.pt}try{if(window.external&&window.external.tran)b+="&tran="+window.external.tran}catch(h){}if(a.b)b+="&"+a.b;var e=a.t,p=e.start,l=[],i=[];for(var c in e)if(!(c=="start"))if(!(c.indexOf("_")==0)){var j= e[c][1];if(j)e[j]&&i.push(c+"."+k(a,c,e[j][0]));else p&&l.push(c+"."+k(a,c))}delete e.start;if(d)for(var m in d)b+="&"+m+"="+d[m];var n=[f?f:"http://csi.gstatic.com/csi","?v=3","&s="+(window.jstiming.sn?window.jstiming.sn:"codesearch")+"&action=",a.name,i.length?"&it="+i.join(",")+b:b,"&rt=",l.join(",")].join(""),g=new Image,o=window.jstiming.c++;window.jstiming.a[o]=g;g.onload=g.onerror=function(){delete window.jstiming.a[o]};g.src=n;g=null;return n}}; </script></body></html> Thanks, Steve

    Read the article

  • Hyper-V RedHat Guest OS

    - by user8160
    I installed RedHat as a guest, and it installed fine. However, I do not have a connection to the internet, I bring up a browser and it cannot surf anywhere. Would someone be able to give me some things to check/do to figure out the issue?

    Read the article

  • Recover strategy single bad sector in moricon

    - by Damon
    This week, my harddisk made me an early christmas present in the form of a single defect sector. To make up for the puny size of the present, it chose a sector inside moricons.dll for that. This means that now the system takes about 5 minutes to boot before Windows gives up and moves on, and there's 2 dozen scary "critical failure" entries in the system log after every boot, which is annoying. OK, admittedly, I shouldn't complain, it could be worse, the bad sector could be in ntldr... SMART info more or less indicates (for what SMART can indicate anyway) that the drive is mostly OK. Soft Read Error Rate has a score of 96, and Current Pending Sector Count has a raw value of 8, which translates to a score of 100. Acronis DriveMonitor makes this an issue (lowering the overall rating to 75%), HDD Health calls it "excellent", giving an overall rating of 95% (which is what this harddisk from day one). No single score is below 95 (power on hours and spin up count), and most are 100 anyway. Well, whatever, I've seen drives with perfect SMART values fail from one second to the other, and drives with moderate values work for years. So, I'm inclined not to put too much weight into that overall. TL;DR Now... to the problem: I don't feel like trashing the disk just yet (that's planned with a new OS install upgrading to Win7 early next year, independently of this issue), but in the mean time, I would still like to have a smoothly running system again. Therefore, I feel tempted to tamper with it, but before I render my system entirely unusable (since I've never done this before), I'd like to verify that my planned procedere is likely to suceed in having a working system again: Copy moricons.dl_ from the Windows install disk, rename it to moricons.zip, and unzip it. This gives an intact 5.1.2600.2180 version (the broken one is 5.1.2600.5512 - but I guess this makes not much of a difference, since it's an icon-only DLL, and an outdated copy should work better than one that can't be read) Run chkdsk /r /f` which will "repair" the file (i.e. delete the file without asking, tell the drive to remap the sector, and toss some unreadable junk into a file with a hexadecimal number) Hopefully Windows still boots after this (is that a reasonable expectation, or do I need to have something like BartPE ready? -- but then again, what's that good for in case chkdsk has nuked the entire file system...) Delete the junk file generated by chkdsk, copy the new DLL to %windir%\system32 Reboot. Pray. Maybe I just shouldn't touch anything, since it still kind of works... if annoying, but it works. Unsure... But, is there anything fundamentally wrong with the planned approach? Is this a sensible approach at all?

    Read the article

  • "Discovering items".... yawn

    - by Kez
    I have a rather annoying issue with my Windows 7 computer whereby it seems to take ages to delete, move or copy folders, even when they are empty. The problem is intermittent, to make matters worse. It just says "Discovering items..." for anything up to a couple of minutes before eventually doing what its told. Any suggestions on where to look to stop this happening? I have tried disabling antivirus realtime scanning to rule that out as the root cause.

    Read the article

  • Opening a non-currupt PST file

    - by Kyle Brandt
    I have a a PST file that a user is unable to open. The permissions look fine to me, and I had to repair the file, but now the PST file is repaired and I can open it from my Outlook (In other words, seems to be a local issue). Any recommendations on what to do to see why the user can't open it with their outlook?

    Read the article

  • IIS server unable to access the file path

    - by anil namde
    I have installed web application behind IIS which access the files/directory located on the another machine. Now while developing the application using the Visual Studio 2008 then Developer Web Server it provides when build the site was able to access the location. However when the site is deployed on the IIS its saying Unable to access the file because of insufficient permissions. Can some one please help regarding what could be the issue ? any idea or pointers could be helpful. Thanks,

    Read the article

  • IIS server unable to access the file path

    - by anil namde
    I have installed web application behind IIS which access the files/directory located on the another machine. Now while developing the application using the Visual Studio 2008 then Developer Web Server it provides when build the site was able to access the location. However when the site is deployed on the IIS its saying Unable to access the file because of insufficient permissions. Can some one please help regarding what could be the issue ? any idea or pointers could be helpful. Thanks,

    Read the article

  • Citrix printing permissions

    - by ANervousTwitch
    Im trying to get some users printing from a remote citrix server at another company. It seems as though they require local administrator privileges to do so, however Ive been instructed not to give them that. They can print normally from their desktop without issue. Is there some special permission to allow citrix to print through to local/network printers short of granting local admin to these users?

    Read the article

  • Grizzly server - works with IP, but not with domain name

    - by Hitchhiker
    I'm hosting a grizzly web service on a Windows 7 Pro machine (embedded in a regular Java process), and it is binding to http://my-domain-name. When trying to hit the service from another machine, requests to http://my-domain-name fail (fiddler shows error code 502), but requests to http://my-ip work. When the service runs on a Windows Server 2008 machine, this doesn't happen (both requests succeed). What could be the issue?

    Read the article

  • How to log size of cookies in request header with apache

    - by chrisst
    We have an issue on our site with cookies growing too large. We have already expanded the acceptable header size and throttled the cookie sizes for now, but I'd like to figure out what the average client's header sizes are, specifically of the cookies. I've created an apache log that captures the cookies being set on each request: LogFormat "%{Cookie}i" cookies But this just spits out the entire contents of all cookies in the header. Is there a way to have apache just log the size (or just length of the string) per request?

    Read the article

  • Only 6to4 addresses are accessible via 6to4 interface

    - by wRAR
    I have a configured 6to4 interface. I can communicate with any other 6to4 host (2002::/4), but not with other hosts, packets just seem dropped. I've tried several configurations and OSes so I think this is a provider issue. How can I make sure it is it and what should I tell my provider?

    Read the article

  • Running psexec against Window 7?

    - by Jimmy C
    I can't seem to get psexec commands to work on Windows 7 Enterprise. I'm trying to execute a command like this: psexec.exe \\172.16.165.157 -u jc -p testing123 -c -f -i cmd.exe But, I keep getting the error: PsExec v1.94 - Execute processes remotely Copyright (C) 2001-2008 Mark Russinovich Sysinternals - www.sysinternals.com Error copying C:\WINDOWS\system32\cmd.exe to remote system: Access is denied. I'm trying to execute the command from a machine running Windows XP Professional. Has anybody seen an issue like this?

    Read the article

  • Pasted image hides behind text even when set to be inline

    - by John
    I copied an image from MSPaint and pasted it into a Word document I'm working on. For some reason the image hides behind the text even with the default "in line with text" setting. Trying other settings don't work as expected either. It does the same when I insert a picture from a file. Can anyone shed any light what would be causing this and how to fix it - I am guessing some formatting issue in the existing document?

    Read the article

  • Trac changesets not displaying

    - by Robin Weston
    I have setup Trac to integrate with our VisualSVN server. It all seems to work fine, except when I am viewing changesets, it only shows the files that have changed, not the actual diff of what has changed in the file. In addition, when I am viewing a particular file in Trac, the 'diff' link is not visible, which I have seen in other installations. I assume it is probably a configuration issue, or maybe the way the data is being entered in the first place. Any suggestions?

    Read the article

  • How to show the "Home" folder for SSRS 2008 in SQL Management Studio

    - by kd7iwp
    I just installed SSRS 2008 on a development server and when I connect to it with SQL Management Studio I do not have all of the items in Object Explorer that I had with our SSRS 2005 installation. For example, we are missing the "Home" folder which lists all the reports. When we browse to our /ReportServer URL we can view and run our reports, so I know they are on the server. Does anyone know if this is a permissions issue or something else?

    Read the article

  • is ROBOCOPY a good form of backup as a tool?

    - by dasko
    seeing how it does not do a write verify unless scripted (or is it needed?) is it a decent option to dump a few folders to another server? i am just worried about whether the data, after being copied, might be corrupted but you would not be aware of it being buggy? i have used it in the past without issue but am seeking feedback in case i missed something obvious.

    Read the article

  • What happens to missed writes after a zpool clear?

    - by Kevin
    I am trying to understand ZFS' behaviour under a specific condition, but the documentation is not very explicit about this so I'm left guessing. Suppose we have a zpool with redundancy. Take the following sequence of events: A problem arises in the connection between device D and the server. This causes a large number of failures and ZFS therefore faults the device, putting the pool in degraded state. While the pool is in degraded state, the pool is mutated (data is written and/or changed.) The connectivity issue is physically repaired such that device D is reliable again. Knowing that most data on D is valid, and not wanting to stress the pool with a resilver needlessly, the admin instead runs zpool clear pool D. This is indicated by Oracle's documentation as the appropriate action where the fault was due to a transient problem that has been corrected. I've read that zpool clear only clears the error counter, and restores the device to online status. However, this is a bit troubling, because if that's all it does, it will leave the pool in an inconsistent state! This is because mutations in step 2 will not have been successfully written to D. Instead, D will reflect the state of the pool prior to the connectivity failure. This is of course not the normative state for a zpool and could lead to hard data loss upon failure of another device - however, the pool status will not reflect this issue! I would at least assume based on ZFS' robust integrity mechanisms that an attempt to read the mutated data from D would catch the mistakes and repair them. However, this raises two problems: Reads are not guaranteed to hit all mutations unless a scrub is done; and Once ZFS does hit the mutated data, it (I'm guessing) might fault the drive again because it would appear to ZFS to be corrupting data, since it doesn't remember the previous write failures. Theoretically, ZFS could circumvent this problem by keeping track of mutations that occur during a degraded state, and writing them back to D when it's cleared. For some reason I suspect that's not what happens, though. I'm hoping someone with intimate knowledge of ZFS can shed some light on this aspect.

    Read the article

< Previous Page | 283 284 285 286 287 288 289 290 291 292 293 294  | Next Page >