Search Results

Search found 7625 results on 305 pages for 'scraper sites'.

Page 143/305 | < Previous Page | 139 140 141 142 143 144 145 146 147 148 149 150  | Next Page >

  • why my system gets frozen?

    - by user93019
    Is now something like 2 weeks since my system gets frozen while browsing the web, using apps, watching videos on web sites, basically using any part of my netbook, even when i'm only using one program (like the web browser) this happens. Is so frozen, that the only option i have left to do is to PRESS and hold the power button, nothing else works; I have seeing an error message telling me about a "compiz" error, but not all the times, and this happens every day, some times 2 or 3 times, during the same day. Why does this happens? how can i fix it? please help. Just to let you know i have my system up to date on 12.04 32 bits version, working on an Acer aspire AOD257, 1 Gb RAM, Intel® Atom™ CPU N570 @ 1.66GHz × 4 .

    Read the article

  • IIS and content caching

    - by JayC
    I'm a web developer and administer of a Windows 2008R2 Could Instance with IIS 7. I recently made an update to our website, but when I revisited the website, the website was being viewed with old stylings. I did a refresh (shift + reload button in Firefox) and of course the website displayed as it should. I didn't worry about it, until my client had the same issue in Safari. So, my question, in general, is, how do I prevent this from happening again, and yet still afford some caching of our site? I noticed we did not have content expiration set up on our webserver sites, so I've set that up, but did I really need to? I've also looked at Etags, and, honestly, it's hard for me to know whether or not I should use them or not. One comment I read somewhere there isn't really any issue with Etags scenarios in IIS (even in webfarms)... but, I dunno. Anybody have any suggestions, links, info? Thanks.

    Read the article

  • How to calculate maximum number of request in 128 MB VPS performance?

    - by ifdion
    I am a newbie here, please let me know if I'm using wrong webmaster terms. I am currently setting up a VPS for a multi site WordPress. The VPS uses Debian 6 LNMP setup and the DNS is being taken care by another service. Currently the VPS is running non multi site WordPress with -+ 83 MB RAM out of 128MB. As far as I know the performance is relative to the number of request, not the number of sites in the multi site setup. The question How do I calculate maximum number of request in with that setup? If the information is not enough, what other factor do I need to know? Thank you in advance.

    Read the article

  • How should I handle search engines auto-correcting the spelling of a site's name?

    - by Nathan G.
    A client's site and company is called 'Tranin Communications' (Tranin is her last name). It ranks well in searches for her name but rather poorly in searches for the name of her site/company. I realized that this is largely due to* search engines (Google especially) assuming that the query was misspelled and automatically including results for both 'train communications' and 'communications training'. Both of those queries yield many high-ranking sites that completely drown out hers. Sometimes Google even shows results for 'communications training' instead of 'tranin communications', hiding her site altogether. Is there a way to report an incorrect auto-correction to Google or something I can do to discourage this behavior (e.g. a meta tag)? My searches have come up cold, any suggestions would be appreciated. *I've come to this conclusion because her site ranks very highly when the same queries are put in quotes.

    Read the article

  • Windows Azure : le dev camp arrive, venez découvrir la page dédiée a l'évènement sur Developpez !

    Le 20 juin aura lieu la journée Dev Camp consacrée à Azure. [IMG]http://i.msdn.microsoft.com/hh868108.azure-camps(fr-fr,MSDN.10).png[/IMG] Cette journée est l'occasion de découvrir tous les services Cloud d'Azure (SQL Azure, Stockage avec Windows Azure Storage, Back-end, etc.), d'apprendre comment réaliser des projets et héberger des applications ? ou des sites webs - sur la plateforme. L'Azure Dev Camp abordera également les applications multi-tiers et la manière de « migrer, intégrer et étendre votre code et vos applications existantes grâce à Windows Azure ». Cette journée abordera aussi la construction d'APIs Web pour enrichir des applications mobiles iOS, Android et bien sûr Windows Phone. Enfin, le rendez-vous...

    Read the article

  • Making sub domain the new main domains for ssl

    - by Dean Legg
    What would be your best advise for changing your main domain to a sub domain? Are site used to be example.co.uk but has now changed to https://secure.example.co.uk/. Any example.co.uk url's re-direct to the new secure domain. Effectively the example.co.uk is now just there to redirect any links and is no longer part of the sites url structure. I have added a new domain to Google Webmaster https://secure.example.co.uk/ and added the site map. Waiting for it to be indexed. Is there anything else you would advise and will this take away a lot of the juice from all the links I developed for example.co.uk? Guessing this is not best practise as I have struggled to find any information online based on this subject.

    Read the article

  • Domain in PENDINGDELETE, question about drop

    - by kcdwayne
    A domain I want is in the pendingDelete stage according to WHOIS. I have been monitoring it since redemptionPeriod, and it entered into pendingDelete 5 days ago today. After checking a few services (SnapNames, etc), they report it is scheduled to drop on the 11th (7 days, by my calculations). I'm not quite sure what to believe. The domain isn't highly valuable. It is to me and one other company. I can see no backorders placed on the big name sites, so I'm thinking of trying to get it without a backorder service. Any insight as to when it will actually drop? I've read 11AM-2PM PST, but I'm unsure. Thanks.

    Read the article

  • how to install gerix on ubuntu 12.04 using backtrack repositories?

    - by werever
    I got instructions from here, and I found several sites with similar instructions but somethig is wrong with the repositories on 12.04. This should work for 11.10 and previous versions. Any ideas how to install backtrack repos and some back track apps on 12.04? I read about downloading the .deb package for gerix and manually making the same folder structure that gerix uses on backtrack, then manually run the python script, but this way didn't work for me either. How can I get gerix installed and working on my Ubuntu 12.04 system?

    Read the article

  • How to report abuse to website hosting company (GoDaddy)

    - by lgratian
    I'm not sure if this is the right place to ask such a question... Let's say that a website posted a picture of me, without my consent, and I want to be removed (it's something private, could compromise my career if it's seen by someone that shouldn't). I sent them an email asking nicely that they should remove it, but they didn't respond and the picture is still there. Using 'Whois' I found that the website is hosted by GoDaddy. Is there a way (an email address, for ex.) to report to GoDaddy that one of the sites they're hosting does something illegal and to force them to remove the photo? I searched the site and found nothing about such a thing. Thnaks in advance!

    Read the article

  • Geographic location settings

    - by JochemTheSchoolKid
    I am building an website. It has an .nl domain. Now only my domain is showing up on google.nl I hope I can change this somehow that it could be findable in all google's (like google.com / co.uk) and so on. If I look on google forums. They say go to webmaster tools and change your geographic position over there. But I have added this site and I am not able to change it there because there is no select box. I dont have any idea were to search (yes I searched on google offcourse) or where to ask for this special problem. So maybe here can someone redirect me or explain me what is possible and what not. The question is can I make an .nl domain findable in (almost) all google search sites? And so on how can I do that. Picture of my google webmaster tools (nl): http://i.stack.imgur.com/ZuP4L.png

    Read the article

  • Alternatives to JS in the browser: your experience [closed]

    - by Andrea
    In these days there are many projects whose aim is to bring new languages to the browser by compiling them to JavaScript. Among the others one can mention ClojureScript, CoffeScript, Dart, haXe, Emscripten, Amber Smalltalk. I understand that in part the response to this answer depends on the particular project. But, anyway: Have you tried any of them? Are these tentatives suitable for production sites? Should I want to try one of those, what are the downsides I should carefully evaluate? By the way, I expect the answer to be slightly misleading for CoffeeScript, since its semantics is very close to that of JavaScript, so I guess it should be easier to make it work seamlessly with the existing JavaScript ecosystem.

    Read the article

  • How should I study a competitor's off page SEO?

    - by Chris Adragna
    What are the things I need to do, and with what tools to know what a competitor has working for him/her off-page (free and paid tools -- please suggest both)? First of all, I'm supposing I want to see all of the sites linking in and see what anchor text is used. Is there something that would report on the anchor text linking in, such as counting the keyword phrases used as anchor text? Also, it would be helpful to see where the PR is coming from by PR, such as, listing inbound links by PR of the page linking in. Lastly, if I'm missing something, here in the way of off-page attributes, please say so.

    Read the article

  • Do they ask too much on this job?

    - by user58404
    I am looking for web developer job and this job description caught my eyes. I am not sure how much they offer but I was wondering if anyone here meets all of their requirements? To me, that's a lot of knowledge. 2 to 4+ years experience building web sites and applications in a professional environment Strong working knowledge of HTML5 and CSS3 Strong working knowledge of JavaScript, jQuery, AJAX Working knowledge of Ruby on Rails or similar MVC framework Working knowledge of ExpressionEngine, Wordpress or similar CMS Experience administering a LAMP-based server Experience with cross-platform and cross-browser website testing Comfortable working with version control (preferably Git) Proficient with Adobe Photoshop, Illustrator, and Fireworks Comfortable working on a Mac Self-starter with excellent time-management skills with the ability to meet challenging deadlines Ability to work independently with minimal supervision Desire to work on a small team Bonus Skills: Experience deploying to Heroku or similar PaaS provider. Experience developing Facebook applications A strong sense of design Cool open source projects (send us your Github account!) Advanced working knowledge of server administration and website deployment. Java and/or .NET experience

    Read the article

  • Website with over 1 million posts with not much textual content

    - by Far Se
    I've made a website which crawls files from all over the Internet and I feel like Google will ban me if I sent it sitemaps which contain all of these pages (1m+), because they contain only the file name/size/no of downloads and the download link(s). I'm considering this thought because I've made another website like this in the past and Google banned me after one week with the reason: "spam", even it was not (maybe somebody falsely reported me?!). Does someone have an idea about how to keep Google form banning my website? I've seen several other sites like mine and they don't get banned or... anything. And also, should I sent the sitemap or wait until Google indexes every page as it finds them? Thanks in advance :)

    Read the article

  • How do I run Microsoft Explorer 8 plus on Ubuntu?

    - by user169487
    First off, you should know I'm a Ubuntu newbie who has become passionate since my strangely low-priced Dell computer showed up with Ubuntu loaded on it and it could do almost everything I need from a computer. But now I've got a problem. I know, it's like asking how to use a cassette player on your computer, but in South Korea EVERY website is made to work with exclusively on MS Explorer PLUS Active X, which Microsoft finally killed with 9. Hence, I have to use 8, as there are some Korean sites I need to access for searching, banking, etc. So I see that the way to go is to load PlayOnLinux, but then I it tells me to delete these. Do I really have to do that? Should I just hit "Install Anyway"? Or would that be just asking for trouble?

    Read the article

  • How Can I Disable Windows 7's Aero Performance Warnings?

    - by Jason Fitzpatrick
    You know your computer isn’t cutting edge, but there’s no need for Windows 7 to constantly remind you. Read on to see how you can disable its constant nagging to adjust your color scheme to improve performance. Today’s Question & Answer session comes to us courtesy of SuperUser—a subdivision of Stack Exchange, a community-drive grouping of Q&A web sites. Secure Yourself by Using Two-Step Verification on These 16 Web Services How to Fix a Stuck Pixel on an LCD Monitor How to Factory Reset Your Android Phone or Tablet When It Won’t Boot

    Read the article

  • Reader Poll: Are You Switching to Internet Explorer 9?

    - by Jason Fitzpatrick
    The recent release of Internet Explorer 9 polished up the IE line with a new interface, hardware acceleration, and more. Is enough to convince you to switch? Yesterday we gave you a full run down of the Internet Explorer 9 release and highlighted the major improvements including: a totally new interface that’s even more streamlined than that of Google Chrome, direct access to your video card for super fast acceleration, customizable menus for pinned sites, HTML 5 support, and a combined address and search bar, among other improvements. Is it enough to sway you? Cast your vote in the poll below and then sound off in the comments. How-To Geek Polls require Javascript. Please Click Here to View the Poll. How to Create an Easy Pixel Art Avatar in Photoshop or GIMPInternet Explorer 9 Released: Here’s What You Need To KnowHTG Explains: How Does Email Work?

    Read the article

  • How to pass information across domains to ask for newsletter only once?

    - by Michal Stefanow
    Lets assume following scenario, I have two sites: example1.com example2.com When user visits 1 there is a prompt "please signup to a newsletter". Same thing happens when user visits 2. However when navigating from 1 to 2 I don't want signup form to be shown. My first thought were 3rd-party cookies, but it seems that they are blocked / not working: http://stackoverflow.com/questions/4701922/how-does-facebook-set-cross-domain-cookies-for-iframes-on-canvas-pages?rq=1 http://stackoverflow.com/questions/172223/how-do-i-set-cookies-from-outside-domains-inside-iframes-in-safari?rq=1 Another thought is to append #noshow for each URL but that would require some work - for instance a script that would intercept click / tap events and modify URL structure depending on the address. (but that seems hacky) I wonder if you know a robust well-established solution to this issue? Thanks

    Read the article

  • BleachBit: How to Completely Clear URL History in Firefox?

    - by tSquirrel
    14.04 / Firefox 29.0 I've been using Bleachbit to clear usage/file history, and for the most part it works great. However, it doesn't seem to clear the website hostnames out of the URL, at all. These addresses are not bookmarked. Also, the total URL isn't preserved, just the hostname. Visit site http://www.bluesnews.com/some_random_URL_string Exit Firefox Run Bleachbit, with ALL Firefox options selected Restart Firefox Check history: completely empty, other than bookmarked sites. www.bluesnews is NOT bookmarked Type "blue" which is Firefox automatically completes as "http://www.bluesnews.com/" Alternate Step #3: Use Firefox's built-in "Clear History" and select ALL entries with a time frame of "Everything". Same result as above. My inquiry in BB forums hasn't been responded to. I found Dan's proposed solution, however changing autocomplete in about:config only turns off the function, it doesn't actually stop storing URLs.

    Read the article

  • Why Is the Localhost IP 127.0.0.1?

    - by Jason Fitzpatrick
    Geeks the world over know their local host as 127.0.0.1, but why is that specific address, of all available addresses, reserved for the local host? Read on to delve into the history of local hosts. Image by GMPhoenix; available as wallpaper here. Today’s Question & Answer session comes to us courtesy of SuperUser—a subdivision of Stack Exchange, a community-drive grouping of Q&A web sites. 7 Ways To Free Up Hard Disk Space On Windows HTG Explains: How System Restore Works in Windows HTG Explains: How Antivirus Software Works

    Read the article

  • Best practice, or generally best way to set up web-hosting server, permissions, etc. [closed]

    - by Jagot
    Hi, I'm about to set up a server upon which a friend and I will be hosting web sites, and I'll be using Debian. I've set up a LAMP solution many times just to using for local testing purposes, but never for actual production use. I was wondering what are the best practices are in terms of setting the server up, in reference specifically to accessing the web root directory. A couple of the options I have seen: Set up a single user account on the server for us both to use and use a virtual host to point to the somewhere in the home directory, e.g. /home/webdev/www. Set each of us up a user account, and grant permissions in some way to /var/www (What would be the best way? Set up a new group?) I want to get this right when I first set this up as there won't be any going back for a while once our first site is up and running. Appreciate any guidance in advance.

    Read the article

  • Can anyone recommend a good SMS gateway for PHP?

    - by Nick
    I'm looking for an SMS gateway that has the following criteria: Supports sending from short codes Opt-in capabilities for end-user (will require user to opt in before sending message) Has a clean API, preferably an XML/ REST API, although SOAP is fine The cheaper the better Honestly, I keep searching and searching all permutations of "SMS Gateway API" etc... on Google and I can't find any great services that can support all of these criteria. There are a lot of really SEO optimized sites and I can't tell which of them are legit. I'm looking for something that will be easy and quick to implement, and none of my colleagues have any recommendations. What APIs have you used (preferably with PHP) that have gotten this job done? Which services have been the most pleasant to deal with, both technically and support-wise? Update: Will be sending primarily (probably solely) in the US.

    Read the article

  • Does setting document.domain via script interfere with Google Analytics?

    - by Seth Petry-Johnson
    I have a site, www.example.com, that displays some secure content from forms.example.com in iframes. To enable cross-frame navigation, pages on both sites use JavaScript to set the document.domain to just "example.com". I am using Google Analytics on www.example.com, but the GA site is not showing any data. It indicates that the tracking code is found (the status icon is a green checkmark), but no data is reported. The GA profile lists the website as "www.example.com". Is this a supported scenario? Is my script interfering with the GA code in some way?

    Read the article

  • How does delicious.com avoid being sued for copyright infringement?

    - by Stanish
    With the recent redesign of delicious.com, they've added a much more graphical home page. The site continues to be a service for people to bookmark and share websites they come across on the web. The delicious home is now made up of images taken from those linked sites. See for yourself at http://delicious.com I would like to know what in the law allows them to do this, considering the images represent the main content of the page, and they clearly do not own copyright to those images? I know there is some leeway given to search engines where it is considered fair use to use a small portion of the content if the aim is to lead people to the originating site. Does that apply here?

    Read the article

  • Qualifying 'happiness' in software development?

    - by mummey
    It occurred to me today that often the real goal of questions asked on sites such as this one (where the questions tend to be more open-ended than say, SO) is for the OP to become happier upon achieving the result. We often excuse this by saying our desire is to be more productive or release a better product, but if you continue to look down this path you can determine that the OP seeks greater productivity or product-quality because those are important to his/her 'happiness'. With that in-mind I ask this: Have their been efforts to study software development from this perspective? In other words, what practises increase happiness in those who develop software as a career, and who, if anyone, has researched this specifically? As I mentioned above, they may include strategies that increase productivity or improve product quality, but by no means should they be limited to just those.

    Read the article

< Previous Page | 139 140 141 142 143 144 145 146 147 148 149 150  | Next Page >