Search Results

Search found 9721 results on 389 pages for 'quicktest pro'.

Page 247/389 | < Previous Page | 243 244 245 246 247 248 249 250 251 252 253 254  | Next Page >

  • What is the reason why some websites are hacked? [closed]

    - by adietan63
    I just want to know. What is the reason why some website are hacked? Is it the website itself or is it the web server? I'm so curious about this because i want to develop my website and I just want to know what are the things I need to do to protect my website? Assuming that i will start it from the scratch. Please give me advice or other technical stuff that will open my mind to developed my website that has security features.. Thank you.

    Read the article

  • How can I allow search engines to index my invite only website in ruby on rails?

    - by tstyle
    I have a ruby on rails website that will be in invite-only mode for the next couple of months. Currently I have it set up so visits to any page performs an authentication: before_filter :authenticate, :except => [:beta] //authenticate checks for a logged in user But the webpage has a lot of content that I would like to see indexed by search engines, and I was wondering if there's an easy way to allow crawlers to do their work? I am not very knowledgable on SEO related stuff at all, so sorry if this is an suboptimal way to phrase the question.

    Read the article

  • Blocked Google Engage Account [closed]

    - by fasheikh
    I have a Google Engage account, which keeps showing me: Your Google Account has changed The email address on your Google Account has changed. The account that controls your Engage membership now has an email address of: fasheikh15%[email protected]. How can I fix this? Go to the sign-in page: Sign In Log in with your normal password but instead use the following email address: myemailadd%[email protected] You will then be given the option to switch your Engage member profile to: [email protected]. If you wish to do this then just follow the on-screen instructions when you get to the Switch Accounts page. I've been contacting them since April to no avail so I thought I'd see if anyone here has any fixes?

    Read the article

  • pages still show up in google search even after disallowed in robots.txt [duplicate]

    - by Jota Onasys
    This question already has an answer here: With Robots.txt disallow all, why was my site still getting traffic? 5 answers Why is it that some pages still show up in google search even though disallowed in robots.txt? Is the best solution here to remove the Disallow from Robots.txt and just add noindex, nofollow meta tag to those pages you want blocked? Or should I submit a request to Google directly to remove those pages?

    Read the article

  • How to track inbound HTTP traffic using Plesk 10.4.4?

    - by hypercrypt
    I am running Plesk 10.4.4 on a Debian 6.0 server. The outbound traffic is being tracked but the inbound http traffic seems to be 0 at all times, i.e. looking at the DomainsTraffic and ClientsTraffic the http_in column is always 0. Is this a setting that I have missed? I've had a look and cannot find anything. How do I get Plesk to track the inbound HTTP traffic? I have already made sure that Home Tools & Settings Server settings had 'Include in the traffic calculation' set to 'inbound and outbound traffic', yet this does not solve the problem. Apache allows inbound traffic to be logged using %I in the log format, is there a way to get Plesk to add the %I to the log and then use that in bandwidth calculations?

    Read the article

  • Which is the best way to catch an expiring domain name? [closed]

    - by newspeak
    I know a similar question has been asked, but I really don't know what to do. There is this .com domain which is currently on redemption period and should likely be available again within a month. I was wondering which is the best way to get it at a reasonable price. I don't think it's a highly valuable domain, it shows to have very bad ranking and has 0 exact same searches according to adwords. Why it is valuable to me is very simple: I have a project responding to this name. I already own the .net domain and would love to have the .com. I discovered the domain was going to be available thanks to an email I received by a backorder site. I did some research and these guys have a bad reputation on the web. I did further research and found that more reputable (at least in theory) companies should be the likes of snapnames, pool, namejet, godaddy, etc. I am a bit suspicious using these drop cathing services: What if they shill bids? What if they make it go into auction even if I'm the only person interested? What if I raise attention and interest to the domain by backordering? I just would rather wait for it to be deleted and available again to register it manually. It is really not an interesting domain name, and I don't think anyone would care to have it. But what if the domain is already being watched by the domain industry sharks? I did a whois research and my desired domain nameserves point to domcollect.com, which appears to be an auction site. What if I decide to wait for manual registration and I miss the chance to get it? I'm willing to spend the 60/70$ fees these sites require, but not really more than that. Suggestions? Thank you very much. I'm a bit confused and undecided.

    Read the article

  • Unable to debug javascript?

    - by linkme69
    I’m having some problems debugging an encoded javacscript. This script I’m referring to given in this link over here. The encoding here is simple and it works by shifting the unicodes values to whatever Codekey was use during encoding. The code that does the decoding is given here in plain English below:- <script language="javascript"> function dF(s){ var s1=unescape(s.substr(0,s.length-1)); var t=''; for(i=0;i<s1.length;i++)t+=String.fromCharCode(s1.charCodeAt(i)-s.substr(s.length-1,1)); document.write(unescape(t)); } I’m interested in knowing or understanding the values (e.g s1,t). Like for example when the value of i=0 what values would the following attributes / method would hold s1.charCodeAt(i) and s.substr(s.length-1,1) The reason I’m doing this is to understand as to how a CodeKey function really works. I don’t see anything in the code above which tells it to decode on the basis of codekey value. The only thing I can point in the encoding text is the last character which is set to 1 , 2 ,3 or 4 depending upon the codekey selected during encoding process. One can verify using the link I have given above. However, to debug, I’m using firebug addon with the script running as localhost on my wamp server. I’m able to put a breakpoint on the js using firebug but I’m unable to retrieve any of the user defined parameters or functions I mentioned above. I want to know under this context what would be best way to debug this encoded js.

    Read the article

  • Does Googlebot (and/or search engines) index a forwarded page? [duplicate]

    - by user2889419
    This question already has an answer here: HTTP and HTTPS impacts on SEO 1 answer Let's say I have example.com domain, and I force the user to use the HTTPS over HTTP. The question is as browsers just accept and load the forwarded/new page (when the request for http://example.com - https://example.com), does the Googlebot (or other search engines) accept the forwarded page and index the new page and just ignore the old page? In other word, does search engines accept HTTPS beside the HTTP?

    Read the article

  • How can I make this script output each categories item per category [closed]

    - by Duice352
    Ok so here is the deal currently this script outputs all the products in a parent category as well as the products in the child categories. What i would like to do is seperate the output based on child categories. All the child categories are in the array $children and the string $childs. The parent category is the first array element of $children with the following ones being the actual children. The category names are stored in the database $result as " $cat_name ". I want to first Display the cat_name then the products that fall in that category and then display the next child cat_name and items, ect. Any suggestions of how to manipulate the while loop that cylcles through the rows? <?php $productsPerRow = 3; $productsPerPage = 15; //$productList = getProductList($catId); $children = array_merge(array($catId), getChildCategories(NULL, $catId)); $childs = ' (' . implode(', ', $children) . ')'; $sql = "SELECT pd_id, pd_name, pd_price, pd_thumbnail, pd_qty, c.cat_id, c.cat_name FROM tbl_product pd, tbl_category c WHERE pd.cat_id = c.cat_id AND pd.cat_id IN $childs ORDER BY pd_name"; $result = dbQuery(getPagingQuery($sql, $productsPerPage)); $pagingLink = getPagingLink($sql, $productsPerPage, "c=$catId"); $numProduct = dbNumRows($result); // the product images are arranged in a table. to make sure // each image gets equal space set the cell width here $columnWidth = (int)(100 / $productsPerRow); ?> <p><?php if(isset($_GET['m'])){echo "You must select a model first! After you select your model you can customize your dragster parts.";} ?> </p> <p align="center"><?php echo $pagingLink; ?></p> <table width="100%" border="0" cellspacing="0" cellpadding="20"> <?php if ($numProduct > 0 ) { $i = 0; while ($row = dbFetchAssoc($result)) { extract($row); if ($pd_thumbnail) { $pd_thumbnail = WEB_ROOT . 'images/product/' .$pd_thumbnail; } else { $pd_thumbnail = 'images/no-image-small.png'; } if ($i % $productsPerRow == 0) { echo '<tr>'; } // format how we display the price $pd_price = displayAmount($pd_price); echo "<td width=\"$columnWidth%\" align=\"center\"><a href=\"" . $_SERVER['PHP_SELF'] . "?c=$catId&p=$pd_id" . "\"><img src=\"$pd_thumbnail\" border=\"0\"><br>$pd_name</a><br>Price : $pd_price <br> $cat_id - $cat_name"; // if the product is no longer in stock, tell the customer if ($pd_qty <= 0) { echo "<br>Out Of Stock"; } echo "</td>\r\n"; if ($i % $productsPerRow == $productsPerRow - 1) { echo '</tr>'; } $i += 1; } if ($i % $productsPerRow > 0) { echo '<td colspan="' . ($productsPerRow - ($i % $productsPerRow)) . '">&nbsp;</td>'; }

    Read the article

  • Where is the source of domain search? [closed]

    - by All
    There are several websites providing service of searching for free domains (websites, not registrars). I wonder where is the source of these searches? This search cannot be based on local database, as it needs live data (of available domains). The only possible way (to me) is to fetch every query from the original NIC (e.g. nic.com), but I was unable to find an API for this service. How to find a source to write a script for domain searching?

    Read the article

  • What's the best way to include a hosted blog on a site?

    - by Blankasaurus
    The way I see it you have a couple of options for including a blog on your site. Roll your own Install Something Like Wordpress Use a hosted service like Tumblr Obviously, if you can get away with Option 3 you have done yourself a favor. I have been using Tumblr and pointing blog.mydomain.com to the domains.tumblr.com CNAME to get a free blog that looks like it belongs to my site. I want to be able to display my blog on my root level hostname like mydomain.com/blog. The whole point of moving from blog.mydomain.com to mydomain.com/blog is for SEO purposes. Whats the best way to go about this and still get to leverage a hosted blogging platform like Tumblr?

    Read the article

  • Alternatives to PHP [closed]

    - by kaz
    We are starting a project, which goal is to create new frontend interface to our product. Old version was created in PHP, very poorly written. We are choosing the language and frameworks that we want to use in new version. Requirements: New interface will be communicating with API. Application will not have it's own database. We don't have a big team, 3 max programmers for entire project. The main programmers are PHP veterans and knows some other technologies (Rails, C, C++, some Java) but not in professional level. But overall they are good and experienced programmers. So: We want to find a good alternative to PHP. I like Rails very much, but whole ActiveRecord model will be useless, when using application API. Java needs a lot of configuration and someone who is expert in Java to properly run this project. Also, in Java there are a lot of big and complicated enterprise frameworks - not very good for 2-3 programmers team. Python - I don't know Python and don't know good and experienced programmers who knows PY - but it's not so complicated and big as Java and maybe in long period it's good alternative for PHP. What are your thoughts?

    Read the article

  • help me being the next zuckerberg [closed]

    - by Kraken
    I really expect someone to help me with my problem as stated below. I started learning website making by having tutorials from w3schools.com but after doing taking the HTML tutorials I don't think I really can do anything with it. Maybe I don't know everything about HTML yet. I know the online tutorials are not enough but what I need you guys to tell me is that how can I learn making some nice websites. I have vacations for two months now and I think it would be the best time for me to learn making websites as I really love to be a website developer. Now I would like to start from the scratch. I know some C language (though I know it will not help me). So tell me from where shall I learn HTML and what all things shall I learn after that and in what order. If you can also help me with the material (let it be books or online tutorials). But I would like to know how can I gain confidence that I can make some websites as just learning online tutorials doesn't help me much. I also tried some video tutorials but in that too, they will just tell me some of the functions and not all. So what am I supposed to do.. just know a limited functions as they want me to know in video tutorials or do something else.

    Read the article

  • Is my htaccess setting hurting SEO?

    - by Ramanonos
    I have a site that I have redirecting to https. I do this to leverage wildcard SSL for my password protected pages. Everything seems to work fine with testing. For example, whether you type in http or www, you always get redirected to the SSL https... That said, I have about 200-300 external backlinks -- many high quality, yet google webmaster (along with SEOMoz), shows I have just 4... Huh? I'm embarrassed to say I just discovered this. This has led me to hypothesize that maybe my settings in htaccess is messed up, so google isn't recognizing a link because it's recorded on another site as http, instead of https. Maybe? At any rate, here is my simple htaccess setting for 301 www to http, and from http to https. RewriteCond %{SERVER_PORT} !443 RewriteCond %{HTTP_HOST} ^www\.example\.com$ [NC] RewriteRule ^(.*)$ http://example.com/$1 [L,R=301] RewriteCond %{SERVER_PORT} 443 RewriteCond %{HTTP_HOST} ^www\.example\.com$ [NC] RewriteRule ^(.*)$ https://example.com/$1 [L,R=301] Like I said, everything works fine for redirect over https, so I'd rather not screw up what works. On the other hand something is very wrong with google finding all my back links, so I need to fix something... I'm just wondering that maybe google isn't picking up a my backlinks from other websites recording me as http because I'm at https. Maybe google doesn't care and it's some other issue. Am I barking up the right tree? If so any quick fixes? Thanks as always!

    Read the article

  • VPNs - The ins and outs of IPSec & VPNs in general [on hold]

    - by Magus
    I have to decided to mess around with VPNs on my home router, to access a couple of servers in the back room of my house, however, I went into this thinking happy thoughts and easy-peasy-lemon-squeezy... Now I feel like doing drugs, as if maybe that will help me understand the myriad of terms which come with this nifty little tool. Basically, I do understand WHAT a VPN is, but I have no idea how to set one up. I have a Cisco router ( will supply more info if needed ), and would prefer to use IPSec for this ordeal. I'd like to have the following terms explained ( and yes, I have used the famous Google to help, close but no cigar ) : "Local Secure Group", "Remote Secure Group", "Remote Secure Gateway", how different types of "Keys" work ( again, I know the basics ), and for the "Add VPN Config" screen on the connecting device: "Server, does it have to be an address, or just an IP?", "Account; is this the tunnel name?", I am going to assume 'Password' means the Key, "Group Name; or is THIS the tunnel name?", and "Secret; I halfish know what this is..." I would really appreciate any contribution made, no matter how small, even if it includes a redirect. I just want to learn. Thanks in advance! Magus

    Read the article

  • Good places to submit a business profile?

    - by Rob
    We have a new graphic/web design site that we've created and we'd like to give it a boost in terms of back links. What are the obvious places to submit a website to achieve some good back links? And what would be the recurring work load (if any) that those back links would need? Are there any good industry specific websites we could submit a profile to? E.g Twitter, would need constant interaction and new content with back links to the website. UPDATE: I've always felt it's worth it but after several years trying to submit different websites, I've never successfully managed to get a site into DMOZ!!

    Read the article

  • Will uploading our .docx files on scribd and embedding the files on our website affect search engine rankings?

    - by user1439968
    We have prepared notes for university students which are on .docx format. And we want it to put on our website for viewing. We tried one option. Uploading the files on scribd and embedding it on our website for viewing on scribd viewer. Will making documents available on srcibd viewer on our website affect search engine rankings ? Will search engines treat it as duplicate content as those are already uploaded on scribd and we are embedding it on our website ? On scribd we have set the uploaded documents as 'private' though. And if it affects, can you suggest any suitable way to make .docx files to be viewed on our website that doesn't affect search engine rankings ?

    Read the article

  • Is Google indexing pages that has no connection with other pages? [duplicate]

    - by Grkmksk
    This question already has an answer here: How did Google find my unlinked newly created pages? 3 answers I am working on a web project that has nearly 100 thousand instant users and there is a webpage that we are using for test cases. There are no links pointing to it from other pages. It shouldn't be indexed by Google or any other search engines. "noindex" can be used in this situation, I know but I wonder if Google (or any others) indexes this page, if I don't do anything to prevent it.

    Read the article

  • Content from a domain I used to own is appearing in my twitter feed

    - by user19424
    I had an old domain with a WordPress setup hosted by GoDaddy. I changed the business name and moved everything over to a new site, new domain, and new host a couple of years ago. I let the old domain expire. It was recently purchased by someone for cheap, article spun content. Now, anytime they post a new article, it's automatically posted to my Twitter account. I contacted them to remove this, but given the quality of their spun articles, I doubt they will remove it. Is there any recourse I have through Twitter or the host to get this removed?

    Read the article

  • Google Analytics + External Site Statistics Tracking in one application?

    - by Soleil
    My company is a broker in the real estate industry. As such, we send a lot of our listings to sites like Trulia.com and Zillow.com, among others. These sites direct leads to our realtors, and provide us with reports every month detailing the activity our listings have had on their site-- links back to our website, emails generated, etc. Our Marketing and Advertising departments want to take that information and enter it into a system to keep track of everything in one place, for the purpose of producing comparison reports. I cannot find any externally available product that provides this functionality. I would sincerely like to avoid writing this tool myself. Does anyone know of a tool that could do this? In short, an ideal system would: Imports Google Analytics data via API Imports real estate listing site data via CSV import / manual entry Provides comparison reports based on data Does anyone know of anything pre-made that can do this?

    Read the article

  • Can a whitespace regex character be used to perform a javascript injection? [migrated]

    - by webose
    if I want to validate the input of a <textarea>, and want it to contain, for example, only numerical values, but even want to give users the possibility to insert new lines, I can selected wanted characters with a javascript regex that includes even the whitespace characters. /[0-9\s]/ The question is: do a whitecharacter can be used to perform injections, XSS,even if I think this last option is impossible, or any other type of attack ? thanks

    Read the article

  • What kind of redirect (301 or 302) for an email links tracker?

    - by MaxiWheat
    We are developing an email sending application ("à la" Mailchimp). Hyperlinks inserted by our users, in the emails they want to send, are replaced by a tracking URL on our application (https://ourdomain.com/trackingurl?blablabla) which then redirects the email reader to the original URL our users included in their emails. This allows us to record statistics about link clicks. Until now, we used 301 for those redirections, but we noticed that Google began indexing pages on our application which are in fact redirects to other domains. (The title and snippet in Google results are from the other domain, but the link in green is from our application). We took action by adding those urls to our robots.txt, but Google seems to take forever (months!) before removing them for its index and removing them by hand in Webmaster Tools would take a lot of time since there are lot. I would like to know which kind of HTTP redirect (301 or 302) is best suited for this kind of opreation ? Do you think switching to 302 redirects could improve this situation since we don't really want Google to index redirected links from our clients emails ?

    Read the article

  • Should my web app have its own domain name?

    - by Daniero
    I'm going to get a domain name for my personal web page. It will contain my blog, photos and other personal stuff, plus different web apps and tools that I'm working on. I have put quite a lot of work in one of the web apps, and I think it has great potential. It covers a niche in a way that no other page has done before and I can see great possibilities to expand it. Via links to webstores and ads I think it could even make some money, and if I play it right this could be the place to go when you're into the specific niche that it covers (sorry for being so vague). My question is, would it be better for a (potentially great) web app like this to have its own domain name (nameofmyapp.com) instead of being a subpage of my personal page (mypersonaldomain.com/nameofmyapp)? Is the web app more likely to be found by others, via Google etc, with its own domain name? Could a "subpage" give more attention (visitors) to my personal page? pros and cons?

    Read the article

  • when the page scroll to certain point, how to make a second fixed header appear

    - by MikeSm
    I have a page that will be responsive and I also want to add a header that appears once the visitor scrolls for a bit. The header will supplant the main header in order to be visible as the user travels down the page. I think this was a convention that people have used, and I need some help, as i can't really code it from scratch. Has anyone seen examples or tutorials on this. I've looked but can't come up with it.

    Read the article

  • Exclude pages from search results based on device class (mobile/desktop)

    - by user32224
    We're currently building a new responsive website. While working on the site map, we figured that we don't want to show certain sections on mobile devices. This can be easily done by hiding the navigation parts using CSS/media queries. However, the trouble is that the hidden sites would still show up in search engine results. If a user happens to click on one of these links she might happen to see a badly formatted page as we'd use desktop/tablet only code to show images and video. Is there any way influence the search engines to exclude certain pages if the search is done on a mobile device? Do search engines crawl pages once or with a device specific view twice? Could we set a noindex meta tag for a specific device class?

    Read the article

< Previous Page | 243 244 245 246 247 248 249 250 251 252 253 254  | Next Page >