Search Results

Search found 8013 results on 321 pages for 'clean urls'.

Page 162/321 | < Previous Page | 158 159 160 161 162 163 164 165 166 167 168 169  | Next Page >

  • What is the best way to generate a sitemap?

    - by Zakaria
    Hi everybody, I need to build a sitemap for my website. The url will be "www.example.com/mysitemap.html". I know that there are some tools that generate automatically an XML file that contains the reachable URLs and also improve the SEO. So my questions are: How can I build this HTML page going from the generated XML? Or am I wrong and this kind of HTML page is built manually? If not, how do we integrate the XML and convert it to the website? Thank you very much. Regards.

    Read the article

  • Why does a Windows 7 installation grow so seriously and how to shring it?

    - by Ivan
    I store all my data and the most of applications (those legally available in "portable" versions) on drive D: and only use drive C: for the Windows system and some heavily integrated applications like MS Office, Visual Studio, Adobe Reader, Flash Player etc. When I was using Windows XP, 50 GiB drive C: was more than enough. Now, as I've mitigated to Windows 7, it hardly is. Yesterday as I've checked, 7 GiBs were free on drive C:. Then I've installed fresh Windows updates (which were just some tens of MiBs to download) and checked again: now there are only 2 GiBs free. Where have 5 GiBs gone? PS: Don't be surprised my system installation takes actually that much: I've got Visual Studio and SQL Server with complete offline documentation library, but that doesn't explain where does free space disappear on simple Windows updates. PPS: I use an augmented CCleaner version to clean my PC every day, so there are for sure no temporary internet files of recycle bin trash files taking the place.

    Read the article

  • Data Archiving vs not

    - by Recursion
    For the sake of data integrity, is it wiser to archive your files or just leave them unarchived. No compression is being used. My thinking is that if you leave your files unarchived, if there is some form of corruption it will only hurt a smaller number of files. Though if you archive, lets say all of your documents, if there is even the slightest corruption, the entire archive is unrecoverable. So whats the best way to keep a clean file system, but not be subject to data corruption.

    Read the article

  • Unable to Boot from USB External Hard Drive

    - by Josh Stodola
    I recently upgraded my main development machine to Windows 7. This involved wiping out my primary boot drive (Windows XP 64-bit) and starting clean. Before I wiped it, I did a direct disk-to-disk copy to a big external hard drive I have. While I have been able to migrate most of the necessary files without any problems, I was wanting to boot from it today to check a few settings. I plugged in the hard drive, rebooted, changed the BIOS to boot from USB-HDD first. But, no mattter what I do, it always boots from my primary drive to Windows 7. I do not see any kind of error message or anything. How can I boot to Windows XP 64-bit on this external hard drive?

    Read the article

  • Need help converting apache .htaccess code to lighttpd url_rewrite code

    - by miCRoSCoPiC_eaRthLinG
    Hi All, I have this custom written CMS built on XAMPP. I'm trying to test the same on lighttpd to benchmark the so-called performance gains... but am stuck at the rewrite rules. I use the default rewrite rules of WordPress in my .htaccess which hands all url parsing over to my script as long as there aren't any files or directories same as the URL structure. Here's the code: RewriteBase /somedir/ # Let the Script handle all pretty URLs RewriteCond %{SCRIPT_FILENAME} !-d RewriteCond %{SCRIPT_FILENAME} !-f RewriteRule .* index.php [L] Can anyone please help me translate the same to the lighttpd format? I could only (partially) figure out the last line myself going through the lighty docs ... i.e. url.rewrite-once = ( ".*" => "index.php" ) But that of course won't suffice... All my CMS is doing is giving me 404's. Thanks, miCRoSCoPiC^eaRthLinG

    Read the article

  • Any thoughts on A/B testing in Django based project?

    - by Maddy
    We just now started doing the A/B testing for our Django based project. Can I get some information on best practices or useful insights about this A/B testing. Ideally each new testing page will be differentiated with a single parameter(just like Gmail). mysite.com/?ui=2 should give a different page. So for every view I need to write a decorator to load different templates based on the 'ui' parameter value. And I dont want to hard code any template names in decorators. So how would urls.py url pattern will be?

    Read the article

  • How do I protect myself?

    - by ved
    I was poking around at my work computer this evening and was looking at my timesheets. I noticed that all my timesheets had variables in the URLs and I could figure out the numbering scheme for the pages. Then I got a little curious about SQL injection and thought of trying out adding simple SQL injections like "OR 1=1" etc. to see how protected we really were with our timesheet info. One of these strings yielded a friendly error page saying that an error email was sent to the developer. I am concerned that my ID, and request will be seen by the developer , immediately recognized as SQL injection and will be reported to network security officer as a malicious attempt by an employee to hack the timesheet dB. what is my defense? I am really worried.

    Read the article

  • Pipeline For Downloading and Processing Files In Unix/Linux Environment With Perl

    - by neversaint
    I have a list of files URLS where I want to download them: http://somedomain.com/foo1.gz http://somedomain.com/foo2.gz http://somedomain.com/foo3.gz What I want to do is the following for each file: Download foo1,2.. in parallel with wget and nohup. Every time it complete download process them with myscript.sh What I have is this: #! /usr/bin/perl @files = glob("foo*.gz"); foreach $file (@files) { my $downurls = "http://somedomain.com/".$file; system("nohup wget $file &"); system("./myscript.sh $file >> output.txt"); } The problem is that I can't tell the above pipeline when does the file finish downloading. So now it myscript.sh doesn't get executed properly. What's the right way to achieve this?

    Read the article

  • iPhone: fast hash function for storing web images (url) as files (hashed filenames)

    - by Stefan Klumpp
    What is a fast hash function available for the iPhone to hash web urls (images)? I'd like to store the cached web image as a file with a hash as the filename, because I suppose the raw web url could contain strange characters that could cause problems on the file system. The hash function doesn't need to be cryptographic, but it definitely needs to be fast. Example: Input: http://www.calumetphoto.com/files/iccprofiles/icc-test-image.jpg Output: 3573ed9c4d3a5b093355b2d8a1468509 This was done by using MD5(), but since I don't know much about that topic I don't know if it is overkill (- slow).

    Read the article

  • Why is the servlet request not available in the portlet API

    - by John Smith
    The portlet API deos not provide any reference to the enclosing servlet request and response objects. I know it is not the preferred model of interaction with the user, but it seems draconian to remove all access. I understand that for portlet driven interaction with the user, you want to use portlet URLs, and let the portlet container manage all the complexity. However if you have a number of portlets which are basically showing variants of the same data, it makes sense for them to be able to use the enclosing request to drive the data. We ended up using using a Liferay specific call to get the request, and it all seems to work as we wish. However I do feel the guilt. So my question really is, is there an underlying deep reason for the prohibition, or is it just to enforce the authors view of the API environment?

    Read the article

  • need some jquery if-else statement help

    - by zeemy23
    Hello, the code below is broken, but I'm not sure how. I've definitely made some big assumptions here as a newbie. I'm basically trying to create an if else where imBannerRotater functions on #cast if the variable is true and #pram if it is false. How could I fix this to get that result? The # are URLs. Thanks!-zeem $(document).ready(function(){ if (mmjsRegionName == 'CO') { $("#cast").imBannerRotater({ return_type: 'json', data_map: { image_name: 'name', url_name: 'url' }, image_url: '#', base_path: '#', }); } else { $("#pram").imBannerRotater({ return_type: 'json', data_map: { image_name: 'name', url_name: 'url' }, image_url: '#', base_path: '#', }); });

    Read the article

  • How do you check Driver Verifier logs on Windows 7 after catching a faulty driver?

    - by Wolf
    I kept getting BSODs on a clean install of Windows 7 (plus updates), so I decided to run Driver Verifier. I had to select all drivers, since it didn't catch the culprit when I didn't include Microsoft drivers. I know it is not a hardware problem since everything is working fine on Linux and memtest86+ is not reporting any errors in the RAM (8 GB). This time, it caught the faulty driver and gave me a BSOD telling me so. Using WhoCrashed, I could verify what was the last error message with the parameters and the source. Yet, the source is always the kernel (ntoskrnl.exe) and the bugcheck this time was 0xC4 (0x85, 0xFFFFF9804429AFC0, 0x2, 0x11B948). After searching on the web, I found out "the driver called MmMapLockedPages without having locked down the MDL pages." As I am not developing any driver, this is of no use to me. However, I would like to know which driver caused Driver Verifier to trigger an alert, so I can either disable it, or rollback to a previous version in order not to get crashes anymore.

    Read the article

  • Removing emails from all mailbox with certain text in subject

    - by Doug Luxem
    So, we had an errant program kick off about 15,000 emails to our users. I have our spam filters now blocking these on the edge until that gets resolved, but now I would like to clean our our users' mailboxes for them. Is there anyway with Exchange 2007 to remove all emails with certain text in the subject line? It would need to be a partial match because the subject changes slightly for each email. I am trying to do this with Forefront's manual scan job + content filter, but it does not seem to like partial matches on the subject.

    Read the article

  • Removing emails from all mailboxes with certain text in subject

    - by Doug Luxem
    So, we had an errant program kick off about 15,000 emails to our users. I have our spam filters now blocking these on the edge until that gets resolved, but now I would like to clean our our users' mailboxes for them. Is there anyway with Exchange 2007 to remove all emails with certain text in the subject line? It would need to be a partial match because the subject changes slightly for each email. I am trying to do this with Forefront's manual scan job + content filter, but it does not seem to like partial matches on the subject.

    Read the article

  • Prevent query string manipulation by adding a hash?

    - by saille
    To protect a web application from query string manipulation, I was considering adding a query string parameter to every url which stores a SHA1 hash of all the other query string parameters & values, then validating against the hash on every request. Does this method provide strong protection against user manipulation of query string values? Are there any other downsides/side-effects to doing this? I am not particularly concerned about the 'ugly' urls for this private web application. Url's will still be 'bookmarkable' as the hash will always be the same for the same query string arguments. This is an ASP.NET application.

    Read the article

  • Actionscript problems with social share encoding

    - by Rittmeyer
    Hi, I'm trying to make some "social share" buttons at my site, but the urls I generate just don't get decoded by this services. One example, for twitter: private function twitter(e:Event):void { var message:String = "Message with special chars âõáà"; var url:String = "http://www.twitter.com/home?status="; var link:URLRequest = new URLRequest( url + escape(message) ); } But when twitter opens up, the message is: Message with special chars %E2%F5%E1%E0 Something similar is happening with Facebook and Orkut (but these two hide the special chars). Someone know why is this happening?

    Read the article

  • Avoiding a NullReferenceException

    - by Nikhil K
    I have used this code for extracting urls from web page.But in the line of 'foreach' it is showing Object reference not set to an instance of an object exception. What is the problem? how can i correct that? WebClient client = new WebClient(); string url = "http://www.google.co.in/search?hl=en&q=java&start=10&sa=N"; string source = client.DownloadString(url); HtmlDocument doc = new HtmlDocument(); doc.LoadHtml(source); foreach (HtmlNode link in doc.DocumentNode.SelectNodes("//a[@href and @rel='nofollow']")) { Console.WriteLine(link.Attributes["href"].Value); }

    Read the article

  • new ubuntu server installation - java server not receiving packets

    - by James
    I have developed a Java server which listens for UDP packets which come from mobile devices. At the moment, it works correctly if I run my Server on my windows PC. I can receive and respond to the UDP packets sent from my phone (over 3g, not wifi). However, I am trying to run the Server on a different computer which I found lying around my house. I have basically done a clean install of Ubuntu Server edition, updated it, installed the java run time environment and configured port forwarding to allow connections on the appropriate port. I have also tried to disable the firewall. The problem is, it doesn't appear as if the server is receiving any packets and I've exhausted my brain thinking on how to solve/debug the problem. Can anyone suggest to me what the problem might be? or more likely something I can do to help me find out more information about what is going wrong.

    Read the article

  • ASP.NET How expensive is it to call an Application Variable many times?

    - by Matias Nino
    The short of it is: Is it costly to check an Application Variable such as Application("WebAppName") more 10-20 times each time a page loads? Background: (feel free to critique) Some includes in my site contain many links and images which cannot use relative urls due to their inclusion in different paths. Hence these includes contain frequent instances of <img src="<%=Application("Webroot")%>images\image.gif"> Is it expensive to keep calling an Application variable like this? Should I just put the Application value in some local variable to use where needed? IMPORTANT NOTE: I need my webapp to run fine on a server whether it be in the root web ("/") or in a virtual subweb ("/app"). Thanks in advance for any wisdom shared.

    Read the article

  • How To Remove A Broken Program From the Programs and Features List In Windows 7?

    - by Itai
    On a Windows 7 64-bit desktop, there is a program on the Programs and Features list (Control Panel\All Control Panel Items\Programs and Features) that refused to uninstall. The error obtained is: Error 1723. There is a problem with this Installer package. A DLL required for this install to complete could not be run I suspect filer were deleted or corrupted. The big problem is that the program refused to install itself over the broken version. So how to I manually clean up the list of installed program?

    Read the article

  • What are the main differences between: Seaside vs Aida vs Iliad

    - by elviejo
    What are the differences between the three Smalltalk web application frameworks? Some starting points: What is the sweet spot for each framework? in Which case would you use one or the other? What are their weaknesses? Which one has the cleanest URLs? How do they handle Ajax? Do they have some preference in their use of persistence? I'm just trying to decide which framework is appropriate for each kind of application.

    Read the article

  • Multiple Copies of Windows Calculator

    - by Brian Boatright
    Just did a clean install of Win7 x64. I have a Microsoft Ergo Keyboard 4000 and use the calculator key a lot. Previously I could hit it and get multiple copies of calculator to popup. Now it will only show one copy of calculator. I tried adding a shortcut to the calculator app but it has the same limitation. However if I click the calculator icon it will open a new one each time. How can I fix this so each time I press the calculator key it will open a new copy?

    Read the article

  • Generate MVC HTml.ActionLink programatically

    - by swetha
    I have a scenario where I should redirect my Legacy Url to my current SEO Url. and I have 300+ different legacy urls so I am handling all those in my 404 Handler. I have my Legacy Url and the actual Action, controller and parameters it is supposed to use. But the issue is there is already a SEO friendly route for these Action and Controller. so I want to redirect this legacy URL to that friendly route. So I am planning to generate a link like what Html.ActionLink does and want to move permanently. Can anyone help me in generating this link???

    Read the article

  • Why is hibernation still used?

    - by Moses
    I've never quite understood the original purpose of the Hibernation power state in Windows. I understand how it works, what processes take place, and what happens when you boot back up from Hibernate, but I've never truly understood why it's used. With today's technology, most notably with SSDs, RAM and CPUs becoming faster and faster, a cold boot on a clean/efficient Windows installation can be pretty fast (for some people, mere seconds from pushing the power button). Standby is even faster, sometimes instantaneous. Even SATA drives from 5-6 years ago can accomplish these fast boot times. Hibernation seems pointless to me when modern technology is considered, but perhaps there are applications that I'm not considering. What was the original purpose behind hibernation, and why do people still use it? Edit: I rescind my comment about hibernation being obsolete, as it obviously has very practical applications to laptops and mobile PCs, considering the power restrictions. I was mostly referring to hibernation being used on a desktop.

    Read the article

  • Installing win xp media center 2005

    - by wdizz
    I have a Dell 710 XPS H2C running Vista Business, I wish to clean install Win XP media center 2005. I wiped my hard drive with Wipe Drive Pro and proceeded to install Win XP, the install starts and goes thru the various prompts but when it comes to 'installing windows' I get a blue screen that flashes briefly (to fast to get the error code) and then reboots. This happens with any XP system disk (iv'e tried 4), but it installs Vista, 7 or Linux with no problems. I contacted Dell Support who have yet to give me any solutions to resolve the problem. Is there anyone out there with any ideas please.

    Read the article

< Previous Page | 158 159 160 161 162 163 164 165 166 167 168 169  | Next Page >