Search Results

Search found 5380 results on 216 pages for 'webmasters'.

Page 198/216 | < Previous Page | 194 195 196 197 198 199 200 201 202 203 204 205  | Next Page >

  • Is it possible to grant access to a folder in a SVN server using SVN's API?

    - by Splendonia
    I need to develop a web application (Using any language but I'm familiar with Frameworks Symfony2 and Rails), that is able to grant access to a user to a determined folder on another server on the same network from the application's front-end. I found out that SVN has an API and that I could interact with it with PHP or Ruby (Apparently), although I would be willing to program the application on another language, the server where the files are stored is using Windows and I thought on using Virtual SVN server, however I can't find any function on the API to grant users access to files and/or folders or access of any kind, like you usually do using the GUI (VirtualSVN on Windows). Am I missing anything? Is this even possible?

    Read the article

  • Standalone URL 301 Redirect Manager [on hold]

    - by Lex
    I'm looking for a script with a simple interface that helps me manage a large list of 301 redirects of the form: http://example.com/short ---> http://example.com/long-and-descriptive For example the following Wordpress plugin does this job, but it seems excessive to install Wordpress just for this one simple functionality. It looks like my question is similar to this closed question, but hopefully rephrased in such a way that makes it more relevant and "constructive".

    Read the article

  • Somehow Google considers a properly 301'd URL as 200 and is still indexing the new content in old page?

    - by user2178914
    We redirected all the old URL's to new ones properly using htaccess. The problem is Google, somehow is still finding content in the old page(which it shouldn't) and stores it in the cache rather than the new URL. For eg: Old Page- http://www.natures-energies.com/iching.htm New Page- http://www.natures-energies.com/index.php?option=com_content&view=article&id=760 If you type the old URL into the browser it redirects If you fetch the old URL as Googlebot in the webmaster tools the header says 301/permanently redirected. If I try to crawl as any other bot it still says 301 redirected. Even if you click the old link in Google it redirects to the new URL. Only in its cache it shows the old URL and moreover it shows the new content in it! I am stumped on how Google manages to grab the new content and puts in the old URL instead of the new one! One more interesting thing is that if I try a cache for the new page it shows the cache of the new content with old URL! Any help would be appreciated. I am at end of my wits. I think i have tried almost everything. Is there anything that I'm missing to see? You can use this search to find the old url's. Maybe you'll some patterns that i missed. site:www.natures-energies.com inurl:htm -inurl:https|index

    Read the article

  • With Google DFP (Small Business) is it possible to disable AdSense in an Ad Slot on a per-request basis?

    - by Daniel Pehrson
    Setup: I run a network of websites that target different hobby niches and have a section dedicated to community classifieds. I serve advertising on these sites through Google DFP for Small Business with AdSense enabled on the slots. Problem: One of the next sites in my network will be targeting the firearms/shooting industry and as such the classifieds section will not comply with the prohibited content guidelines of AdSense regarding the sale (or coordination of sale) of weapons. I work very hard to comply with the guidelines of my partners even if I don't understand/agree with them and after talking with many people have decided that the best option is to disable AdSense serving on that section of that website, while leaving it on for the rest of the network. Solution: Right now my only idea for this is to duplicate all my site's ad slots and tack a "_sensitive" onto the end of each one (eg. header and header_sensitive) conditionally registering ad slots based on whether or not I am in the sensitive section of the sensitive site. My hope however is that there may be a way to accomplish this without duplicating all my ad slots possibly with some sort of options to the GA_googleFillSlot() call that allows me to say "load ads from this slot but do not serve AdSense no matter what."

    Read the article

  • Serving WordPress menu links in only HTTPS or HTTP depending on how it's accessed

    - by Gelatin
    I have a WordPress site which uses WordPress HTTPS to enable SSL when users access it via that protocol. However, currently the menu links point back to the HTTP version. I want users to be linked to HTTPS pages while accessing the site over HTTPS, but not when accessing it over HTTP. Is this possible? Note: I have tried changing the menu options to use // and / for the links, but in both cases they are just rendered as HTTP links.

    Read the article

  • I think there is a problem with my url encoding

    - by TheGateKeeper
    I took someone's advice and started encoding my image sources, but Google doesn't seem to be able to decode them. I probably did something wrong, because basically Google is taking the full path as the image's name. See this page as an example. If you go on the top most thumbnail and do "Save as", you will see the path is not being decoded. Should I stop decoding or am I doing it wrong? Should I encode only the image name itself? Thanks!

    Read the article

  • Restricting URL w.r.t HTTP method and setting different authentication mechanism for each

    - by user31745
    I shall start with an example. I want to restrict to POST requests only for http://path/to/logical/abc.xml and restrict to GET only for http://path/to/logical/def.xml. How do I put constraints like this as the paths are logical and location directive is not supported in .htaccess? The actual problem is to set different authentication type(basic, digest) on diff logical file.for eg. for abc.xml I want to authenticate for Basic type of authentication and def.xml with digest.

    Read the article

  • mysql database cannot connect with cpanel [closed]

    - by Rafee
    And this question was asked on http://stackoverflow.com/questions/8182119/mysql-database-cannot-connect-with-cpanel <?php $con1 = mysql_connect("mywebsiteip","mysql_username","mysql_user_password"); if(!$con1) { die ("Could not connect " . mysql_error()); } else { echo "Good connection"; } mysql_close($con1); ?> When i run it, it cannot connect to mysql database over cpanel. and i even tried up $con1 = mysql_connect("mywebsiteip:portnumber","mysql_username","mysql_user_password"); Can any let me know, which one is good way.

    Read the article

  • Is it possible to tell a search engine not to index a specific section of an HTML page? [closed]

    - by Justin
    Possible Duplicate: Preventing robots from crawling specific part of a page I know you can use robots.txt to ignore entire pages or sections of your site, but is there a way to tell cralwers like the Googlebot to ignore specific sections of an HTML page? I found this blog post that discusses one method, but it appears only to work for the Google Search Appliance, not the Googlebot. Is there some method for at least Google for to do this?

    Read the article

  • Has anyone used Genetify for A/B testing

    - by Joshak
    I'm working on a small project that will likely run on Wordpress, I always like to run some split testing to improve conversion rates for various goals. Typically if its a small site that I either don't have a budget for or want to keep it as inexpensive as possible I use Google Website Optimizer if I do have a budget I go with Visual Website Optimizer both are great and affordable, but for fun I was checking out alternatives and found Genetify which is an open source project and has some neat features. In searching around I don't see many people talking about it and wondered if anyone here has used it. If so what do you think about it?

    Read the article

  • Get exact time of each click - AdWords

    - by almo
    I have an AdWords campaign running and want to get a list with all clicks and the exact time the click occured. And I want to do this without having a tracking code on my homepage. So it needs to be via the API (AdWords/Analytics). You might think know that this is not possible. But big Bid Management tools (e.g. Kenshoo) know all details of every click and they don't have tracking code of my website (only on the success page of my contact form). How is that possible? The dimensions in AdWords only let me group the clicks at the hour level.

    Read the article

  • Which MIME type to compress? and what If I omit the `type` attribute from the HTML?

    - by rockyraw
    Per my request, my webhost had turned mod_deflate ON. In my Cpanel I now have an "Optimize Website" button. Inside that menu I could either choose: "Compress all content" or "Compress the specified MIME types" with the following default MIME types: "text/html text/plain text/xml" Which option should I choose and why? If I choose option 2, which types should I add (is there a recommended list with the exact way they should be written)? According to Google recommendations, I have omitted the type="text/css" attributes from all CSS references, as well as the type="text/javascript" attributes from all script references. Would this hinder the "gzipping" process?

    Read the article

  • Google Analitycs says source is Search Engine (100%)

    - by Angel Estrada
    I have a problem with my google analytics code, 'cause it says since 2 months that all traffic of my websites is directly from search engine, i'm really sure that must be a mistake 'cause i have made some email marketing and landing page campaing that could help to create referal traffic, maybe somethink goes wrong with the new version of the analytics or i forgot some settings. please any help is aceptable

    Read the article

  • 301 redirect from a country specific domain

    - by Raj
    I originally started using a .do domain extension for my site, but later realized that this country specific domain would prevent us from appearing in search results for places outside of the Dominican Republic. We started using a .co domain extension and redirected all requests to the new domain using an HTTP 301. The "Crawl Stats" in Google Webmaster Tools shows me that the .co domain is being crawled, but the "Index Status" shows the number of pages indexed at 0. The "Crawl Stats" for the .do domain says that it's being crawled and the "Index Status" shows a number greater than 0. I also set a "Change Of Address" in Google Webmaster Tools to have the .do domain point to the new .co domain. We're still not appearing in search results at all even for very specific strings where I would expect to find us. Am I doing something wrong?

    Read the article

  • Using Ajax #! for Google but site is not being crawled any more

    - by user28231
    We used to have all links in a normal way (thus with parameters?), we changed to Ajax loading cause we installed a music player. Here's an example: Old links: http://stereofox.com/post.php?idPost=5326 New links: http://stereofox.com/post.php#!idPost=5326 The snapshot you can get it by adding _escaped_fragment_= after the ?. All of these have the same content, and this image says what has happend to the site since we changed the linking system.

    Read the article

  • Search Result Organization

    - by Vecta
    I'm creating an AJAX live search on a website I'm working on. Users will select values from a few dropdowns and a list of products will be returned based on what they select. Some possible fields would be: color, model, make, etc. What type of organization of search results do users tend to find most useful? Is it better to lump them all together (alphabatized) or is it more useful to lump them together by make? In the past I've tended to group them by "make" but I'm not concerned that this will continually force some items with a make toward the end of the alphabet always to the bottom of the list. Any tips are greatly appreciated.

    Read the article

  • How Does Domain Know Where Your Web Host Is Located [closed]

    - by icu222much
    Possible Duplicate: How Does Domain Know Where Your Web Host Is Located? I just purchased a domain name from RapidNames, and a hosting plan at JustHost. I was told to enter JustHost's name server (ns1.justhost.com) in my domain name's name server field and wait for 24 hours for the process to be complete. I do not understand how RapidNames can find my account on JustHost's server as I am sure I am not JustHost's only customer. I have read the article How DNS Works that John Conde has posted, but I still do not understand the issue. After reading several other articles, I am beginning to understand how it works, but I would still like someone to confirm if I am correct or not. From my understanding, linking your domain name to your web host is a two step process. First, you need to tell your domain name who your web host is. This is done by providing the two DNS server addresses. Secondly, you need to tell your web host which domain names you own by entering your domain names into the domain name manager. As a result, when someone queries your domain name, they will be forwarded to your web host. The web host will look in their database to match the domain name the account's owner, and then serve the appropriate website. I want to confirm if my understanding of how a domain knows where your web host is located is accurate?

    Read the article

  • Links don't work on new server

    - by Beko
    I've recently moved from server with php4 to php5. Now I cant open content in my page. Links look like this "page.com/news-it-is-rainy-day-1.html". But when I try to open link I just get the archive (news.html). I also use smarty template engine. My rewrite rule in .htacces looks like this: Options +FollowSymlinks RewriteEngine on RewriteRule ^news-(.*)\.html$ news.php?do=read&id=$1 RewriteRule ^news.html$ news.php RewriteRule ^index.html$ index.php AddType application/x-httpd-php .xml

    Read the article

  • How do I create a background image on web page?

    - by kasha
    I am a new designer so hopefully this question isn't too basic! How do I create a background image on a webpage for a programmer? I designed the page in photoshop and I would like to know how to send the background image (the 25% opacity buildings overlay). I would be happy to send the main image (but it is too large and I imagine would slow the site and loading time drastically). here is the link to the design... http://problemio.com/home_page_1_1.pdf

    Read the article

  • XML Parsing Error at 1:1544. Error 4: not well-formed (invalid token)

    - by Steve
    I have installed Joomla 1.5.22 on a new hosting account, which doesn't have a domain yet, so it's public URL is http://cp-013.micron21.com/~annimac/ A message saying: XML Parsing Error at 1:1544. Error 4: not well-formed (invalid token). The source code for this message is: <dl id="system-message"> <dt class="error">Error</dt> <dd class="error message fade"> <ul> <li>XML Parsing Error at 1:1544. Error 4: not well-formed (invalid token)</li> </ul> </dd> </dl> There is nothing in /logs to indicate what the problem is. I have uploaded the following folders from a freshly unzipped copy of Joomla 1.5.22: administrator components includes language\en-GB libraries modules plugins templates\ja_purity xmlrpc and the issue remains. I have no custom or additional plugins, modules, or components installed. If I change templates, the problem remains. What is the problem?

    Read the article

  • Crossbrowser issue - navigation-menu [closed]

    - by aztekk
    I'm having issues with crossbrowser compatibility on my navigationmenu for my site. The issue is that it's not working as expected in MSIE. It bugs out on mouseover. The site is run with wordpress and the theme is called GreenChilli. It's a free theme from MyThemeShop and they don't seam to be very active in resolving free theme issues on their forum. Can someone have a look and see if this is an easy fix, or if I maybe have to abaondon this theme for something else? Site is: http://lamslagen.com

    Read the article

  • Storing data on server [closed]

    - by Maciekp
    1.How am I supposed to store data on server, using not only: databases,text files and images? And how someone could implement storing data in fb's graph api http://developers.facebook.com/docs/reference/api/ , so when I go to: https://graph.facebook.com/19292868552 it shows me such data(how it can be stored? I guess it's not Mysql database) PS. Link to article: http://jayant7k.blogspot.com/2009/05/how-facebook-stores-billions-of-photos.html <- How can concurrent users writing requests be solved(while storing data in text file).

    Read the article

  • google analytics statistics

    - by colmcq
    I am compiling a report for a client using google analytics. I have observed that the client has unusually good page view times (5 mins) and excellent bounce rates (<25%). I need to reference research data that validates my assertion that these figures are excellent compared to an industry standard (the industry is ecommerce and gaming). Can you direct me to any published research data that specifies normal bounce rates and page view times for this industry?

    Read the article

  • How to offer a cookie opt in/out to users?

    - by Darkcat Studios
    I intend to use google analyticts, and as I understand it I will need to offer users the option to opt out of cookies. The question is this: I HATE these constant cookie option boxes, everyone I ask it getting annoyed by them too. Its nice to have the option, but we all know they have been in use for well over a decade. So - how big of a deal do I have to make about the fact that I'm using GoogleAn? can I pop a small link at the bottom of the page, maybe integrate it into "Privacy policies" page, and give people the option to opt out there? This would be very much the "Assume the majority of users don't mind, but at least make the option available" stance. Ironically setting a cookie seems to be the only way I can see to enforce the opt-out! as IP's change.

    Read the article

  • Alternatives to Google Sites for a personal web site?

    - by Oleg2718281828
    I'd like to find a hosting solution for my personal, fairly low-traffic, web site. I think Google Sites would suit me perfectly, but I've heard horror stories about Google's algorithms shutting down all your accounts on a whim, with no right of appeal, so I'm naturally worried, because I don't want to lose my Gmail. Edit I'm talking about cases like this one: I tried contacting somebody at Google support ("Surely they should have a support department, right?" Nope, wrong!) (The victim managed to regain access to his Gmail account when his case went "viral", but he never got an explanation as to why he had been locked out in the first place) Are there good, preferably free, alternatives to Google Sites and what are their PROs and CONs? One requirement is that I should be able to point the DNS (foobar.com) to it.

    Read the article

< Previous Page | 194 195 196 197 198 199 200 201 202 203 204 205  | Next Page >