Search Results

Search found 8013 results on 321 pages for 'clean urls'.

Page 169/321 | < Previous Page | 165 166 167 168 169 170 171 172 173 174 175 176  | Next Page >

  • Windows Explorer not responding

    - by William Barnes
    I am running W7 64bit on an AMD Quad 1.8Ghz processor. I have a 500GB system drive and added a 1TB drive. Windows Explorer does it's "not responding" thing regularly. I read the other posts and ran SCANNOW and it came back clean. I also ran AUTORUNS and disabled everything that was not windows related. No luck. I booted in safe mode and it appears that explorer works, but I don't know why. I did read a post awhile back that it could be a Service program that makes explorer not respond. I forgot to bookmark the site and don't remember how to turn off the SP's without crashing the entire system. Any help or suggestions will be appreciated. I have also run CHKDSK on all drives, and there were no errors found. EDIT: It has been in for about 2 years and the problem started about 4 or 5 months ago.

    Read the article

  • Using DNS entries to determine location

    - by Raphink
    I'm trying to think of a clean way to determine the location of machines (mainly, which datacenter they belong to) based on their network settings. I would like it to be dynamic, and I'm thinking of using special DNS records that would be specific to the DNS server in each datacenter. For example, you could have: root@machine1# dig TXT mysite ... mysite 3600 IN TXT "DC1" ... root@machine2# dig TXT mysite ... mysite 3600 IN TXT "DC2" ... etc. I know that DNS has a special LOC record for location, but it takes coordinates, so it doesn't help in my case. Is there a standard way of addressing this issue, another special type of record for it, or some standard entries in TXT records?

    Read the article

  • MovieClip property - what is the url it goes when I click on it.

    - by cosmorocket
    I have bought a flash template in an online store. There are different buttons on it and some of them link to existing local html files from the pack - index-1.html, index-2.html etc. There are also some buttons that don't go anywhere right now but they are intended to be linked to some urls too. They react with animation on mouse rollover and I suppose they should go to some url if I set it somewhere on click. But there is no documentation about that in the pack. I searched through the Flash source file and all html pages files for something like "index", "geturl", also looked through them manually to realize how it could be made but with no luck. So, my question - is that possible that way or another to make that MovieClip button link to some my page, for example, "contacts.html". Thanks.

    Read the article

  • problems with url and email regex when searching text

    - by Grant Collins
    Hi, I'm having problems with regular expressions that I got from regexlib. I am trying to do a preg_replace() on a some text and want to replace/remove email addresses and URLs (http/https/ftp). The code that I am have is: $sanitiseRegex = array( 'email' => /'^([a-zA-Z0-9_\-\.]+)@((\[[0-9]{1,3}\.[0-9]{1,3}\.[0-9]{1,3}\.)|(([a-zA-Z0-9\-]+\.)+))([a-zA-Z]{2,4}|[0-9]{1,3})(\]?)$/', 'http' => '/^(http|https|ftp)\://[a-zA-Z0-9\-\.]+\.[a-zA-Z]{2,3}(:[a-zA-Z0-9]*)?/?([a-zA-Z0-9\-\._\?\,\'/\\\+&amp;%\$#\=~])*$/', ); $replace = array( 'xxxxx', 'xxxxx' ); $sanitisedText = preg_replace($sanitiseRegex, $replace, $text); However I am getting the following error: Unknown modifier '/' and $sanitisedText is null. Can anyone see the problem with what I am doing or why the regex is failing? Thanks

    Read the article

  • Should I log my website's 404 errors?

    - by Ivan Zlatanov
    I have an ASP.NET website, but this question isn't really about technology, it is rather about practice. Should we log our 404 errors? My reasoning: This is a potential vulnerable point because a simple unfriendly user may fill up your hard drive in no time just by requesting wrong URLs! Some browsers often request resources up front - like for example favicon.ico, even if its not there. This is really annoying. But really I would like to know about a broken link if there exists one in my websites. Should I depend on the URL referrer? The problem with the URL referrer is that I cannot distinguish my internal redirect which may be broken with an unfriendly one from outside. What does the practice suggest?

    Read the article

  • Apache loads any file that begins with the same string as used in url. How to prevent this?

    - by MarshallBananas
    If I point to: mywebsite.com/search and there is a file called search.php or search.html or search.inc.php or search.whatthehell.php in website's directory, Apache will point to that file instead of 404'ing. What is even more annoying is that if I point to: mywebsite.com/search/string?also=whatever Apache will still display any file with filename that begins with "search.". Also, all RewriteRules with patterns containing filenames existing in directory are ignored/useless. I'm using Apache 2 on Mac, unmodified httpd.conf. How do I prevent it from redirecting my urls so freely?

    Read the article

  • Is there a Java equivalent for Ruby on Rails "url_for"?

    - by Oren Ben-Kiki
    I have written something like this pretty easily in C# (string GetUrl(new { controller = "foo", action = "bar", baz = "fnord" }), based on the existing capabilities of the XmlRouteCollection class provided by the ASP.NET MVC framework (why it isn't there out of the box is beyond me; the additional required code was trivial). I am now faced with a JSP project, and I need the same ability: centralize the logic for generating all URLs in one place, based on a list of routing rules. Is there some code somewhere I could reuse/adapt to do this in Java? It seems like a common enough requirement, but google proved surprisingly unhelpful in finding something like this.

    Read the article

  • Possible to change the order of the shortcut locations in windows explorer

    - by Codezilla
    When you open a new explorer window by default it loads from the starting position "Computer" which shows your drives, etc. This is all well and good except when explorer loads it bumps the scroll bar down so that half the favorites are cut off. I can scroll up to get to the ones I want but that gets really old doing that 50 times a day. I do clean out my favorites frequently but more are added all the time and used. In trying to come up with a fix for this problem I thought maybe it would be possible to move the link to "Computer" to be at the very top and the "Favorites" area directly under that but I can't figure out a way to do that. Anyone have an idea? Or maybe another solution I haven't thought of?

    Read the article

  • Rewrite all URL requests to https://www.example.com/$1

    - by xylar
    I have two domains, example.com and example.co.uk, that use the same application on my server. I would like to rewrite the address of the URL depending on what the user types in. The only URLs I want are https://www.example.com and https://www.example.co.uk In my .htaccess file I have the following: # Turn on URL rewriting RewriteEngine On RewriteCond %{HTTP_HOST} ^example\.co.uk$ [NC] RewriteRule ^(.*)$ https://www.example.co.uk/$1 [L,R=301] RewriteCond %{HTTP_HOST} ^example\.com$ [NC] RewriteRule ^(.*)$ https://www.example.com/$1 [L,R=301] If I goto http://www.example.com it doesn't add the https, if I goto http://example.com it does. What is the best way of making the ReWriteCond match the www url?

    Read the article

  • Reducing CPU load to absolute minimum [on hold]

    - by user191338
    I have had a couple of things gone missing I believe stolen in my shared apartment and want to run my laptop constantly with a webcam attached, running webcam surveillance software to record/ take pictures when motion is sensed. Id like to take whatever steps are necessary to be able to run the laptop constantly without the fan coming on, as its quite loud and even though it will be hidden it can be heard. Thus Id like to know what steps I can take to reduce CPU to the bare minimum for the laptop to boot up and run the camera software and send images via ftp / email when necessary. I have windows 7 installed, though I can reinstall it clean. Which are the windows services can I turn off, and more extreme disabling or measures of whatever kind which I can take. The OS would need to run the camera, wifi / networking. Thanks very much for any help.

    Read the article

  • Outlook 2007 attachments wont preview

    - by Slace
    When I get an email with an attachment in Outlook 2007 that I should be able to preview (ie - image, word doc, etc) I get the following error: This file cannot be previewed because of an error in the following previewer: Microsoft Outlook image previewer To open this file in its own program double-click it. (obviously the previewer changes name depending on the type) The computer is a clean install (it's only 5 days old) and I remember I was at one point able to preview within Outlook. I'd really like this feature to be back working, I've tried the usual exit Outlook, restart Windows, etc but to no avail.

    Read the article

  • How can I request local pages in the background of an ASP.NET MVC app?

    - by flipdoubt
    My ASP.NET MVC app needs to run a set of tasks at startup and in the background at a regular interval. I have implemented each task as a controller action and listed the app-relative path to the action in the database. I implemented a TaskRunner process that gets the urls from the database and requests each one at a regular interval using WebRequest.Create, but this throws a UriFormatException. I cannot use this answer or any code that plucks values from HttpContext.Current.Request without getting an HttpException with the message "Request is not available in this context". The Request object is not available because my code uses System.Threading.Timer to do background processing, as recommended here. Here are my questions: Is there really no way to make local web requests within an ASP.NET web app? Is there really no way to dynamically ascertain the root path to the web app even using static dependencies in ASP.NET? I was trying to avoid storing the app's root path in the database (as FogBugz does with its "Maintenance Path"), but is this best option?

    Read the article

  • [iPhone app] Inserting special char un NSString for URL use

    - by Dough
    Hi, I'm using HTTP connection to share data with my JSON server. I use URLs like "MyServlet?param1=value1" and so on... I'm now facing a problem with one of my servlet (I can't change it because some other views are using it) : The servlet is working with a syntax including those symbols "{" and "}". The exact syntax is {(value1_value2)(value3_value4)(value5_value6)}{(value7_value8)(value9_value10)(value11_value12)}{(value13_value14)(value15_value16)(value17_value18)} Values are integers, the problem is only when I use "{" and "}" my UrlConnection returns an error for bad URL. I use this to instantiate my NSString : NSString *myURL = [NSString stringWithFormat:@"http://somesite.com/Servlet?PARAM={(%@)}"]; How can I code those char in my NSString ? Thanks in advance !

    Read the article

  • Free duplicate music finder for Mac.

    - by Jack M.
    I'm trying to clean up an mp3 folder which has a plethora of duplicate files in it due to accidentally dragging my music folder into iTunes and having it re-import songs which were already in the play list. I tried writing a quick Python app to md5 all of the files, and delete exact duplicates. This took out ~2gb of files. Unfortunately, however, this does not work on all of the duplicates because of an iTunes feature. iTunes has changed the ID3 title on some of the duplicate songs, which means the md5 of the entire file is different from the same song with a different ID3 tag. Are there any free applications out there (for the mac) which can compare the data of the actual song (ignoring ID3 tags) and determine if duplicates exist?

    Read the article

  • Best way to store data for Greasemonkey based crawler?

    - by Björn
    I want to crawl a site with Greasemonkey and wonder if there is a better way to temporarily store values than with GM_setValue. What I want to do is crawl my contacts in a social network and extract the Twitter URLs from their profile pages. My current plan is to open each profile in it's own tab, so that it looks more like a normal browsing person (ie css, scrits and images will be loaded by the browser). Then store the Twitter URL with GM_setValue. Once all profile pages have been crawled, create a page using the stored values. I am not so happy with the storage option, though. Maybe there is a better way? I have considered inserting the user profiles into the current page so that I could all process them with the same script instance, but I am not sure if XMLHttpRequest looks indistignuishable from normal user initiated requests.

    Read the article

  • "digg" button and encoded url :S

    - by guest86
    Hi! I wrote a php site (it's still a prototype) and i placed a "Digg" button. Placing the button was easy but.... Official manual says "url has to be encoded". I did that with urlencode(). After urlencode, my url looks like this: http%3A%2F%2Fwww.mysite.com%2Fen%2Fredirect.php%3Fl%3Dhttp%3A%2F%2Fwww.othersite.rs%2FNews%2FWorld%2F227040%2FRusia-Airplane-crashed%26N%3DRusia%3A+Airplane+crashed So far, so good but when i want to submit that url to digg, it is recognized as invalid url: http://www.mysite.com/en/redirect.php?l=http://www.othersite.rs/News/World/227040/Rusia-Airplane-crashed&N=Rusia:+Airplane crashed If i place a "+" between "Airplane" and "crashed" (mere end of a link), then digg recognize it without any problems! Please help, this bizare problem is killing my braincells! P.S. for purpose of this answer urls are changed (nonexisting) because, in original, non-english sites are involved P.S.S. Happy New Year! :)

    Read the article

  • Passing URIs as URL arguments in Drupal

    - by wynz
    I'm running into problems trying to pass absolute URIs as parameters with clean URLs enabled. I've got hook_menu() set up like this: function mymodule_menu() { return array( 'page/%' = array( 'title' = 'DBpedia Display Test', 'page callback' = 'mymodule_dbpedia_display', 'page arguments' = array(1), ), ); } and in the page callback: function mymodule_dbpedia_display($uri) { // Make an HTTP request for this URI // and then render some things return $output; } What I'm hoping to do is somehow pass full URIs (e.g. "http://dbpedia.org/resource/Coffee") to my page callback. I've tried a few things and nothing's worked so far... http://mysite.com/page/http%3A%2F%2Fdbpedia.org%2Fresource%2FCoffee Completely breaks Drupal's rewriting http://mysite.com/page/?uri=http%3A%2F%2Fdbpedia.org%2Fresource%2FCoffee Gives a 404 http://mysite.com/page/http://dbpedia.org/resource/Coffee Returns just "http:", which makes sense I could probably use $_GET to pull out the whole query string, but I guess I'm hoping for a more 'Drupal' solution. Any suggestions?

    Read the article

  • How to transform html anchor <a> to WordML

    - by Monomachus
    Hi, I need to transform an anchor tag to WordML without using relationships. Is it possible? I found the w: anchor property but seems it refers only to internal document anchors and not to links or URLs. <w:hyperlink w:anchor="chapter3"> <w:r> <w:t>Go to Chapter Three</w:t> </w:r> </w:hyperlink> So it would be great if something similar would be possible to do without making an Id in relationship document and than referring this id from w:hyperlink. Anyone knows something like that?

    Read the article

  • java for freebsd

    - by rawman
    I am trying to install java on freebsd 7.X and i have done following thing so far download the tar file from this link and store the file in cd /usr/ports/distfiles/ after that i run cd /usr/ports/java/diablo-jre16 or cd /usr/ports/java/diablo-jdk16 make config install clean and uncheck all the options in config screen now after doing all this now i am getting following message to indicate for amd version even if my server is using intel Core2Quad Q9550 2.83 GHz mail# make install === Vulnerability check disabled, database not found === License check disabled, port has not defined LICENSE === Found saved configuration for diablo-jdk-1.6.0.07.02_12 Because of licensing restrictions, you must fetch the distribution manually. Please access http://www.FreeBSDFoundation.org/cgi-bin/download?download=diablo-caffe-freebsd7-amd64-1.6.0_07-b02.tar.bz2 with a web browser and "Accept" the End User License Agreement for "Caffe Diablo 1.6.0". Please place the downloaded file(s) in /usr/ports/distfiles. * Error code 1 Stop in /usr/ports/java/diablo-jdk16. mail# (as before i was getting error for tzupdater so i downloaded tzupdater-1_3_29-2010f.zip and put it in a cd /usr/ports/distfiles/ folder)

    Read the article

  • What does the the reconstruction process of mdadm do exactly on raid10

    - by Azrael
    I've got a system with 4 disks set up as raid10. All disks are usable, and mdadm all states them with UUUU. Due to a recent system crash, the raid is currently reconstruction the raid as it was marked as "not clean," and a reconstruction process was started. On a closer look smartctl shows problems on one disk: sd 0:0:0:0: [sda] Unhandled sense code sd 0:0:0:0: [sda] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE sd 0:0:0:0: [sda] Sense Key : Medium Error [current] [descriptor] Descriptor sense data with sense descriptors (in hex): 72 03 11 04 00 00 00 0c 00 0a 80 00 00 00 00 00 24 cd 78 d4 sd 0:0:0:0: [sda] Add. Sense: Unrecovered read error - auto reallocate failed sd 0:0:0:0: [sda] CDB: Read(10): 28 00 24 cd 75 1e 00 04 00 00 With a research about the reconstruction process, I only found information concerning raid5 but nothing for raid10. Can I replace this problematic disk during the reconstruction process, or will I kill the raid with this?

    Read the article

  • Automated URL checking from a MySQL table

    - by Rob
    Okay, I have a list of URLs in a MySQL table. I want the script to automatically check each link in the table for 404, and afterward I want it to store whether the URL was 404'd or not, as well as store a time for last checked. Is this even possible to do automatically, even if no one runs the script? ie, no one visits the page for a few days, but even with no one visiting the page, it automatically ran the test. If its possible, how could I go about making a button to do this?

    Read the article

  • CD-R suddenly unreadable?

    - by TheD
    I have a CD-R which has worked fine for a good while. All of a sudden, Windows can no longer read its data. It's not scratched, it's clean and Windows can detect the disk and usage of space on it. But whenever I attempt to access the data stored: Explorer crashes Or after 5-10 mins of trying to read it, it opens up and just shows a desktop.ini file in Explorer. This happens on multiple machines. Any ideas? Is there a way to recover the data? For example - some sector by sector recovery software if any exist for CD's?

    Read the article

  • Windows 2003 SP1 terminal server printers Disappear after reboot - HP laserjet 4240

    - by Alex
    had working PCL6 laserjet 4240 driver. needed to downgrade to PCl5, tried but did not get clean install. tried again and it seemed to work (this is 2003 enterprise terminal server SP2) Have over 40 working laserjets (5, 4000, 4100 and 18 of the 4240) After normal nightly reboot the 18 4240 printers were 'gone'. Worked w/Microsoft who said bad HP driver issues - weird since they work on other terminal servers. downloaded latest version, etc. from HP site and can NOT get to work. As soon as I install, then do a Net Stop Spool and Net Start Spooler the printer is 'gone'. Current workaround is to use HP 4000 PCL5 drivers for all of these 4240 printers.

    Read the article

  • Can you control pinterest's "find image" results?

    - by anthony
    Rather than add Pin It buttons through our site, I would like to simply control what images show up in Pinterest's "Find Image" results if a user decides to pin one of our URLs. As of now, "Find Images" allows the user to scroll through the images it finds on the page so they can select which image to pin. The "found" images start with the first jpg in the html file, I'm assuming (could that be a bad assumption??). On our site, this forces a user to scroll through about 15 navigation and promotion images before arriving at the featured product image. Is there any way to specify this image to show first in those results? Maybe through a meta tag, or by adding a class or id to the element? Without a public Pinterest API, this seems like just guesswork, but I wanted to see if anyone else has run into this, or solved this. Thanks.

    Read the article

  • Can't access internet even though everything is working

    - by entity64
    A friend recently upgraded to a new cable internet connection. The modem connects to the router and various PCs and smartphones from her roommates connect to the router. They don't have any problem accessing the internet. She has Windows 8 and can't access any website (via wifi and ethernet). DNS (UDP) is working, DHCP set up everything correctly, Wifi is working, Trace routes and Pings (ICMP) go through with no problem at all. But neither Dropbox nor Skype nor Spotify nor any browser (all TCP) can access any website. The thing is though, she can connect through the university wifi and via a neighbors wifi. It's just her home connection. No firewalls are running and the computer is clean - no malware. How could it be that only her home connection won't work and others do?

    Read the article

< Previous Page | 165 166 167 168 169 170 171 172 173 174 175 176  | Next Page >