Search Results

Search found 20308 results on 813 pages for 'software robots'.

Page 229/813 | < Previous Page | 225 226 227 228 229 230 231 232 233 234 235 236  | Next Page >

  • Google met à jour reCAPTCHA et teste si vous êtes un humain avant, pendant et après que vous ayez interagi avec le CAPTCHA

    Google met à jour reCAPTCHA et teste si vous êtes un humain avant, pendant et après que vous ayez interagi avec le CAPTCHA Le CAPTCHA (Completely Automated Public Turing test to tell Computers and Humans Apart) est un moyen de sécurité sur le net permettant de différencier un humain des robots ou des spammeurs par la génération à l'écran des caractères dont le but est d'être difficilement déchiffrés par les bots.Plusieurs CAPTCHA irritent la plus part de temps les utilisateurs, les mots générés...

    Read the article

  • 5 Reasons to Use An Instant Site Creator to Build Your Websites

    An Instant Site Creator is a piece of software that will allow you to create an unlimited number of websites with just a few mouse clicks. You can put up a new, professional looking site in minutes instead of hours or days, saving you both time and money. Here are five reasons you might want to use an instant site creator, or website builder software, instead of outsourcing the project or spending long hours doing it yourself

    Read the article

  • How to Evaluate the Best Web Builder Through a Website Building Review

    There is a lot of software that are specifically designed for creating websites. This makes it hard for one to choose which website building software to use. There are website builders which are perfect for e-commerce websites while some are fit for small business owners or those who make websites as a form of hobby. To help you find the right website builder, you can check out some of the website building review available online.

    Read the article

  • What is best practice for search engines when a website is under maintenance?

    - by jamescridland
    I need around a week to transition a heavily data-driven website from one back end to another. During that time I do plan to attempt to keep some pages live, but they won't all work well or look brilliant. Some pages won't work at all. What is the best way to ensure I don't scare Google? Should I hide everything from robots.txt, or mark everything that doesn't work as "503", or are there other things that I should be considering?

    Read the article

  • Identifying which pattern fits better.

    - by Daniel Grillo
    I'm developing a software to program a device. I have some commands like Reset, Read_Version, Read_memory, Write_memory, Erase_memory. Reset and Read_Version are fixed. They don't need parameters. Read_memory and Erase_memory need the same parameters that are Length and Address. Write_memory needs Lenght, Address and Data. For each command, I have the same steps in sequence, that are something like this sendCommand, waitForResponse, treatResponse. I'm having difficulty to identify which pattern should I use. Factory, Template Method, Strategy or other pattern. Edit I'll try to explain better taking in count the given comments and answers. I've already done this software and now I'm trying to refactoring it. I'm trying to use patterns, even if it is not necessary because I'm taking advantage of this little software to learn about some patterns. Despite I think that one (or more) pattern fits here and it could improve my code. When I want to read version of the software of my device, I don't have to assembly the command with parameters. It is fixed. So I have to send it. After wait for response. If there is a response, treat (or parse) it and returns. To read a portion of the memory (maximum of 256 bytes), I have to assembly the command using the parameters Len and Address. So I have to send it. After wait for response. If there is a response, treat (or parse) it and returns. To write a portion in the memory (maximum of 256 bytes), I have to assembly the command using the parameters Len, Address and Data. So I have to send it. After wait for response. If there is a response, treat (or parse) it and returns. I think that I could use Template Method because I have almost the same algorithm for all. But the problem is some commands are fixes, others have 2 or 3 parameters. I think that parameters should be passed on the constructor of the class. But each class will have a constructor overriding the abstract class constructor. Is this a problem for the template method? Should I use other pattern?

    Read the article

  • Can You Replace Your Webmaster With WordPress?

    As websites turn toward blogging for an easy answer to being able to have user generated content without creating it in FrontPage, DreamWeaver or some other web development software the need for on-staff webmasters has lessened. Many companies are turning to blogs, and especially the WordPress content management system which puts the power back in the hands of the business owner to add, delete, and edit content without the need for someone versed in web development software and file transfer protocol (FTP).

    Read the article

  • Database Insider - June 2012 issue

    - by Javier Puerta
    The June issue of the Database Insider newsletter is now available. (Full newsletter here) INFORMATION INDEPTH NEWSLETTERDatabase Insider Edition - June 2012 Oracle #1 in RDBMS Share Gartner released its 2011 worldwide RDBMS market share research based on total software revenues, Market Share: All Software Markets, Worldwide 2011, and Oracle remained first in worldwide RDBMS share in 2011. Read More New Independent Report Endorses Oracle Database Firewall In a new KuppingerCole Product Research Note, Martin Kuppinger concludes that Oracle Database Firewall "should definitely be evaluated and is amongst the recommended products in the database security market segment."Read More Read full newsletter here

    Read the article

  • Can I benefit from links to pages on my site which have a `noindex` meta tag?

    - by Noam
    I'm trying to understand if/how I can benefit from people linking to pages on my site which are with pages that have a noindex meta tag. 2 actions I'm considering to perform: Remove the robots.txt disallow to these pages, to make sure inner links get the propagated link juice. Adding a canonical tag to the most similar page that doesn't have a noindex meta tag Are these valid approaches that might help? Any others I should consider?

    Read the article

  • Keyphrases - How to Use Them

    Keywords and phrases are words which trigger a response from the search engine spiders (mathematical robots that crawl the web looking for new content to index). They are effective if they are tuned into what people type into the search engines at this moment in time, and you can find this out through the Google AdWords Keyword Tool.

    Read the article

  • Novell: 20 chances to reinvent itself

    <b>The Open Road:</b> "Novell, once the king of the software world, is like that. Over the years it has built up a broad portfolio of software (with associated revenue streams) in repeated attempts to regain its glory days. That portfolio now stifles its ability to focus on other areas with the most promise."

    Read the article

  • Special 301: FOSS users. Now we're all Communists and Criminals

    <b>Free Software Magazine:</b> "There seems to be no respite from the predations of Microsoft FUD and the machinations of Big Business. Just when it seemed safe to come out of the closet and admit to being a user of free and open source software without being accused of being a Communist, it appears that we are now criminals too"

    Read the article

  • Project in .Net MVC, and would like to outsource [closed]

    - by Tito
    I have an interesting software project in .Net MVC, and would like to find someone to finish it. Is an e-learning website with some "inovatives" tools. Since at the moment I am working in a company, and don't have so much time, I would like to hire somebody to finish it. Is there any free lancer website for .Net MVC developers that I could find somebody to finish it? Or any place that I can share part of the site's sharings with other software developer ?

    Read the article

  • The Microsoft Elephant in the Open Source Room

    <b>Boycott Novell:</b> "Summary: Assorted new reports about how Microsoft abuses &#8220;open source&#8221; to gain control of it, change its direction and goals, or even to misuse the label to promote proprietary software that harms standards and promotes patenting of software"

    Read the article

  • Is multiple domain names and links from same IP causing poor search engine rankings?

    - by John
    I have an ecommerce website which is not doing so well in Google. I am trying to improve this of course, and am looking at some possibilities for why it isn't doing well. The website has four domain names, all of which have been indexed by Google. A few months ago I applied 301 redirects to any requests for two of the domain names so now it is down to two domain names (one is a .net, the other is a .com.au, the others were .net.au and .com). I prefer to use my main domain name (the .com.au), but one of the names has been around for a long time and has more inbound links. According to a PageRank tool, both are PR2. It is a Classic ASP site and up until recently had a lot of querystring parameters. In the last week or so I added URL rewriting so there is now no parameters for most pages. I don't do 301 redirects from the old URLs but instead I add the META canonical tag indicating the preferred new URL. At the same time I redesigned the site and improved title tags, META descriptions, and H tags but it hasn't been long enough yet for Google to index many of these yet. I also looked at what pages Google has indexed and strangely it has some strange pages in the index, there are a lot of pages which are actual keyword searches (more a bunch of random letters than an actual word). What I mean is that it is as if they had typed in something to search for in my search box - there are no links to pages like this and the only way of getting this is to type something in to the search box). So I added a META robots tag with noindex,nofollow anytime that I render pages like this. Years ago I set up a fake price comparison site which lists all my products and links back to my site. It has a different keyword rich domain name but is on the same server and same IP address. It's a completely different layout but does have the same product categories and product descriptions (although I have stripped formatting out of them so they are not identical except in text). I also have a few blog sites which again are on the same server/IP and all have advertising for the website. My questions are: What should I do with the multiple domains, just use one, or continue with two or more? Should I add 301 redirects, not just the META canonical tag? Any idea about Google indexing my search results page, and did I do the right thing with the META robots tag? Is the fake price comparison site likely to be causing problems? Are all the links to the site from other domain names but the same IP address likely to be causing problems? Thanks for any help. Sorry for so many questions in one.

    Read the article

  • Evolving Architectures Part II but Design is emergent

    This is part II of a series on agile architecture. You can read part I here.In the previous installment I provided a definition for software architecture and raised the apparent friction between the up front design implied by software architecture and the YAGNI approach and deferred requirements prompted by agile development in the large. This [...]...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Turning off the Visual Studio &ldquo;Attach to process&rdquo; security warning&hellip;

    - by Shawn Cicoria
    When you’re urnning under x64 you have to affect 1 addition spot in the registry to disable this warning – which clearly should only be done by folks that know what they’re doing. NOTE: affecting the registry can be harmful – do so at your own risk. Windows Registry Editor Version 5.00 Windows Registry Editor Version 5.00 [HKEY_CURRENT_USER\Software\Microsoft\VisualStudio\10.0\Debugger] "DisableAttachSecurityWarning"=dword:00000001 [HKEY_CURRENT_USER\Software\Wow6432Node\Microsoft\VisualStudio\10.0\Debugger] "DisableAttachSecurityWarning"=dword:00000001

    Read the article

  • Share USB devices between 32 bit and 64 bit systems

    - by Sreejith S
    The latest version of USBDeviceShare, the USB over IP software, supports sharing USB devices between 32 bit and 64 bit editions of Windows. A USB device plugged in to a 64 bit PC running 64 bit version of Windows can be shared and remotely accessed from a 32 bit PC and vice versa. Download USBDeviceShare USB over Network software from http://www.sysnucleus.com/usbshare/usbshare_download.html Keywords : USB over Ethernet, USB Server, Remote USB Access, Share USB

    Read the article

  • Kraft 0.40 for KDE 4 Released

    <b>openSUSE Lizards:</b> "I am very happy to announce the new stable version 0.40 of the KDE office software Kraft. After eleven month of porting work, Kraft 0.40 is the first version which is based on the KDE 4 software platform."

    Read the article

  • The Integrity Challenge

    - by andyleonard
    Introduction This post is the twenty-eighth part of a ramble-rant about the software business. The current posts in this series are: Goodwill, Negative and Positive Visions, Quests, Missions Right, Wrong, and Style Follow Me Balance, Part 1 Balance, Part 2 Definition of a Great Team The 15-Minute Meeting Metaproblems: Drama The Right Question Software is Organic, Part 1 Metaproblem: Terror I Don't Work On My Car A Turning Point Human Doings Everything Changes Getting It Right The First Time One-Time...(read more)

    Read the article

  • The Job of SEO Spiders

    The World Wide Web, also known as the Internet, is a very complex world. Search engines like Google need a software program that can read what's on the web. The said software program is known as bot or spider or crawler.

    Read the article

  • Which license text to sell my application

    - by ZedTuX
    I would like to sell (for a low price) my new application without DRM (like does Machinarium) on the Ubuntu software center and on a website with Paypal. When I have registered my application the Ubuntu software center and I've selected the licence, I just found "Proprietary" but in my application I have to provide a licence text that something different than just "Proprietary", isn't it? Do you know where I can find this kind of licence text ?

    Read the article

  • What Are Search Engine Crawlers?

    People don't know how they get relevant results for their search queries on a search engine. Most of them believe that these websites were submitted to the search engine. Few others think that there is some software tool that is searching for the relevant websites. Robots and spiders are the software tools that keep on searching the web to find new pages.

    Read the article

  • SEO - Coaching Newbies

    All search engines use algorithms and each search engines have different ones. An algorithm is the formula that the search engine uses to evaluate your web pages. The robots will crawl all pages on your site but not all pages will be indexed.

    Read the article

< Previous Page | 225 226 227 228 229 230 231 232 233 234 235 236  | Next Page >