Search Results

Search found 9721 results on 389 pages for 'quicktest pro'.

Page 243/389 | < Previous Page | 239 240 241 242 243 244 245 246 247 248 249 250  | Next Page >

  • What is the most secure environment for multiple CMS sites? [closed]

    - by Brian Gulino
    I wish to run about 50 Joomla or WordPress low-traffic websites on 1 server, or part of a server. Each website will be managed by its own, naive owner who will have be able to access the Joomla or Wordpress backend of the website. I am concerned about security and isolation as my users will periodically get into trouble by not protecting their sites properly. Two alternatives I know of exist: Run one Linux system with multiple websites under Apache. Follow current Joomla and WordPress security tips. Increase the isolation of the individual sites by using mpm-itk, which will allow each website to run as its own user. The alternative to this is to run virtualization software such as the Xen hypervisor. Each site would have its own, virtual Linux system. I lack the experience needed to make this decision and I am asking which path to take. Obviously, there may be other alternatives that I haven't considered.

    Read the article

  • What is the best taxonomy from Google's perspective?

    - by ZakGottlieb
    I was wondering what the best way is to structure a new website in Google's eyes. Currently, it contains two top-level categories (X & Y), and clicking a term under either one will result in the URL: www.nameofsite.com/X/X type term, or /Y/Y type term Technically, it is correct to group all "X type terms" under X and "Y type terms" under Y, but we could probably be more granular and break all articles into 5-6 top-level categories by breaking Y up into more specific categories. Given that the current URL structure will eventually result in 1000's of "X type terms" and "Y type terms" under just two top-level categories, would it be more advisable to have several of these, as suggested? Thank you in advance.

    Read the article

  • Ok to use table for calculator? [closed]

    - by max
    I'm a php/mysql guy, and have been trying to brush up on my frontend skills. So this weekend I made a four function calculator in javascript. But when I started to work on the presentation, I found myself adding extraneous markup just to achieve what a table tag naturally does. Just so we're on the same page, this is the intended layout: 789+ 456- 123x c0=/ It it possible to generate this grid using neither a table, nor extraneous markup? Thank you.

    Read the article

  • Change player in javascript game [migrated]

    - by KLUSTER
    Game: onClick startbutton mathrandom for first player who starts the game. 4 Pictures: 2 of it player1 and player2. another 2 Player turn. need help: on button click next player turn function game(){ var PlayerTurn; PlayerTurn=parseInt(Math.random()*2); if(PlayerTurn==0){PlayerTurn=1;window.document.player1.src="Cache/Player3.PNG";} else{PlayerTurn=0;window.document.player2.src="Cache/Player4.PNG";} } Any help is appreciated.

    Read the article

  • Can't Make Changes via FTP on my Ubuntu Server

    - by Rev
    I'm very new to server management, I literally just set this up for the first time a couple weeks ago, no idea what I'm doing. In order to allow WordPress to run updates, apparently everything needs to be owned by www-data, but if everything is owned by www-data, I can't FTP with the user revxx14. When I try, I can't make any changes (deletions, additions, updating files, nothing. I get permission denied errors across the board). Is there a way to give my user the same permissions as www-data, so that I'll be able to keep www-data the owner, but still be able to make changes via FTP? Thanks.

    Read the article

  • Backlinks for nonexisting page

    - by Michal
    I've bought domain, which was previously used by somebody back in 2007. Now I've realized that internet is full of backlinks that point to non-existing parts (pages) under my website-domain (for example to mypage.com/whatever, where whatever is not present on my website, so 404 error shows). I want to ask, are these links counted by google (for pagerank) and other search engines, or not. So do I have to redirect these links to existing pages in order to be counted?

    Read the article

  • How can I prevent my site from being branded a "content farm?"

    - by Fredashay
    I'm building a small social Q&A site. Another Q&A site that I use was recently branded by Google as a content farm and removed from Google results. I know what Wikipedia says is the definition of a content farm (low quality paid articles and spammy text across the page to catch search engines). That other site I use doesn't do those things, so there must be more to it than that. I want to make sure I don't do anything that causes Google to think my Q&A site is a content farm. What should I do, or avoid doing in designing my site layout?

    Read the article

  • curl to itself behind firewall

    - by xtreaming
    I have a server A which is configured behind a firewall and has 30.x.x.x public adress and 172.x.x.x internal address. I'm trying to make a php Curl call from a script located on that server, to the 30.x.x.x external IP of that server but the curl call cannot be resolved. It seems that server A does not have a route to that IP. Did you encounter any similar situations? Any chance to solve it through static routes?

    Read the article

  • How to prevent a search engines from indexing a section of a page?

    - by BrunoLM
    I have many pages with lots of text in it. But I will always have two sections of text and I want to prevent one section from appearing in search results, the other section must be indexed. <p class="please-index-me">text</p> <p class="get-out">never index me please</p> I thought that maybe if I load the "please don't index me text" with Javascript maybe search engines wouldn't look for it. But I am not sure it would work and this is not really nice. I was wondering if there is a way to tell search engines "hey, this text you can't grab, move on". So, is there a way to do it?

    Read the article

  • alternatives for google adwords? [on hold]

    - by cauchy
    I'm making tests on my landing page, and trying to get some traffic using google adwords. Everytime I make a change or add a new keyword, I have to wait for two or three days for google to aprove the keywords. I'm in Italy, don't know if this happens anywhere else. But is really upsetting. Note: they keywords are things like fortran, SaaS, python... So, I'm looking for an adwords alternative that is fast. That is, if I change the keywords, I can start getting traffic as soon as possible without the long approval time. PD: I've found many alternatives to adwords, but they don't say anything about how fast your adds and keywords are approved which is my major problem right now. That, and price.

    Read the article

  • The Best Title for my Skill Set [closed]

    - by nickelpickle
    I'm about to branch off into the freelance world. I'm starting an LLC and I'd like some input on what I should title myself as the owner. For example "creative specialist" or "creative technician" or something like that. My services would be: Website design / development Graphic design: icon design, templates, web graphics, business cards / brochures / letterheads / etc. Writing: content writing/copywriting, technical writing, editing / proofreading / copyediting Photography, photo editing Does anybody have any ideas on some general terms that would apply to this type of business?

    Read the article

  • The best way to snatch an expiring domain?

    - by SilvrSun
    There's a domain that I've been looking to acquire that is expiring on the 30th of this month. I don't think it is very popular, and the guy hasn't seemed to update the website in two years now. So, I was doing some research and came across this site that seems to review some "snatching" services, but the article is quite outdated. So, I'm wondering if anyone can offer any newer information on the topic, or whether the recommend any services for helping me acquire the site in question?

    Read the article

  • What are the minimum steps that I should follow to ensure that my web site is accessible to the disabled?

    - by Tim Post
    I am trying to follow a very important standard that I must admit I have ignored up until recently. I want to make sure that my pages are accessible to a large portion of people that have disabilities. I focus mainly on tutorials that are text and image intensive, but no video / flash or any kind of animations. What is a checklist that I can follow to ensure that many people with disabilities can have a good experience when using my web site, and what disabilities should I be most conscious of? I know that I can't possibly please everyone. I have gone through the W3C guidelines, however I'm not entirely sure what standards apply to me. I'm not building web applications, I'm building mostly wiki like information exchanges, blogs and the occasional forum.

    Read the article

  • Does any CMS natively support something like aquaBrowser/Vufind

    - by nus
    What I'm looking for is to set up a CMS website with a tag cloud/search system where when you click a tag it shows as a search filter, and you get a new tag cloud which only shows tags from the articles that have the primary tag. This should let the user easily include or exclude tags from the search. Also i would like to let the user filter on publication date by using a slider... Check this for an example of how aquabrowser does this (just search for something and play with the tag cloud): http://kidscatalog.columbuslibrary.org It's not perfectly what I want and I don't want to use flash, but I like the concept alot. If something like this does not exist and i might have to code it myself, which cms is most recommended (eg easy to extend, already has tag clouds and search filters,...)

    Read the article

  • Why would urls submitted in google webmaster tools drop to 0?

    - by ambient
    Why would urls submitted in google webmaster tools drop to 0? It's a small site, only like 20 pages, I submitted the xml sitemap and for about a week it said 20 urls submitted. A day or so ago it indexed about 17 of the pages, but today when looking it not only says that 0 are indexed but also 0 have been submitted. I did a site search on google and found clearly that pages are indexed, is this just an error on google webmaster tools? Any help or thoughts would be appreciated. Thanks!

    Read the article

  • Since Google reduces the value of links alongside nofollow links, what is an alternative?

    - by SharkTheDark
    Since 2009, Google counts nofollow links also as outgoing links, and thus reduces the value of the other links. What are some alternatives to stop Google counting outside links from my page? If I make links appear on my page source like this: <span hrefs="http://link" rel="nofollow" link="true">Link Name</span> and then in JavaScript replace span with a tag and replace hrefs with href for every span tag that has link="true". Will this help?

    Read the article

  • Site loads Extremely slowly over HTTPS; loads perfectly over HTTP - Sudden and new issue

    - by guest234239048
    My business' website suddenly started loading extremely slowly over HTTPS last night and continues through today. The facts: Page loadtime via HTTPS - 2 minutes +, HTTP - 3 seconds max. NO updates were done on any site files in the past 2 weeks. This is on Shared Hosting HTTPS has worked perfectly over the past 9 months then suddenly failed yesterday. Hosting company and SSL issuer say there are no problems on their end. - Searches for others having similar issues return no results, it appears to be just me... Site is primarily run via php/mysql Currently attempted troubleshooting: Tried all major browsers and different versions - same result. Tried 2 separate ISPs - same result. Tried proxies - same result. Tried 3 separate computers - same result. I'm basically at a total loss here. Does anyone know what could cause such a thing to happen? Please help guide troubleshooting.

    Read the article

  • Is this the place to ask SEO questions [on hold]

    - by user39206
    I'm adding business listings to an existing website of which Google has sort of been referencing well but my Meta tags have a lot to be desired! Due to other work commitments I have kind of neglected the site but now think I should put the proper time and effort into getting it just right as it does earn me money. I am a developer so I have all the necessary skills to build the site, it's almost done!. Now I'm just a little worried that I could do something wrong and loose the rapport I've built up with Google in the past or be ban. Especially with the new tags I see like: <link rel=”publisher” href=”https://plus.google.com/[YOUR BUSINESS G+ PROFILE HERE]“/> Really I just want to know for now, is this the place to ask SEO question?

    Read the article

  • "Mega Menus" for SEO [duplicate]

    - by Thought Space Designs
    This question already has an answer here: How do I handle having to many links on a webpage because of my menu 4 answers I'm using the term "Mega Menus" loosely here. I'm redesigning my WordPress site (it's going to be responsive), and as part of the redesign, I was debating incorporating some sort of descriptive menu setup. For example, normal navigation drop down menus come in the form of unordered lists of links like so: <nav> <ul> <li> <a href="#">Link1</a> </li> <li> <a href="#">Link2</a> </li> <li> <a href="#">Link3</a> <ul> <li> <a href="#">Sub Link1</a> </li> <li> <a href="#">Sub Link2</a> </li> <li> <a href="#">Sub Link3</a> </li> </ul> </li> <li> <a href="#">Link4</a> </li> </ul> </nav> What I'm looking to do is build my drop down menus with more information than your standard menu. For example, I have a top level link named "Team", and under that link, I want to make a large drop down that contains head shots, headers (in the form of styled p tags) and brief (<100 words) descriptions of each team member (only 2 currently). I want to accompany this with a "Read More" link that takes you to their actual team page. This is just one example, of course, and the other top level links would also have descriptive drop downs in the same fashion. On mobile, I was planning on hiding the "mega menu", and delivering a standard unordered list of links. Here's what I was thinking for overall structure and syntax: <nav> <ul> <li> <a href="#">Home</a> </li> <li> <a href="#">About</a> </li> <li> <a href="#">Team</a> <ul> <!-- DESKTOP --> <li class="mega-menu row"> <a class="col-sm-6" href="#"> <div class="row"> <div class="col-sm-4"> <img src="#" alt="Team Member 1" /> </div> <div class="col-sm-8"> <p class="header">Team Member 1</p> <p>Short description goes here.</p> </div> </div> </a> <a class="col-sm-6" href="#"> <!-- OTHER TEAM MEMBER INFO --> </a> </li> <!-- END DESKTOP --> <!-- MOBILE --> <li> <a href="#">Team Member 1</a> </li> <li> <a href="#">Team Member 2</a> </li> <!-- END MOBILE --> </ul> </li> <li> <a href="#">Contact</a> </li> </ul> </nav> Can anybody think of any potential SEO ramifications of doing this? I'm not going to be loading these menus full of links, so it shouldn't hurt page rank, but what are the effects of having a good bit of text and maybe even forms within nav elements? Is there such a thing as overloading nav with HTML? EDIT: Here's an example of what the menu would look like rendered on desktop. I'm currently hovering the "Team" menu, but you can't see because my mouse went away when I took the screenshot. EDIT 2: This question is not a duplicate. I'm not going to have "too many" links in my menus. I'm wondering how having images and text inside of header navigation will affect my menus. Also, I don't just want "yes, this is bad" answers. Please cite your sources and be specific with reasoning.

    Read the article

  • Impact of migrating home page from HTTP to HTTPS on search results

    - by 2Stroke SEO
    I've had to change one of my site's home page to HTTPS from HTTP. I had plenty of links coming into the HTTP page and was performing well in Google against many of my targeted search phrases. I did a 301 redirect from the HTTP page to transfer the link juice to the HTTPS page (and to prevent duplicate content issues) but my search rankings have tanked which indicates no link juice has been transferred. My PageRank has vanished - which I'd expect - but I'm really surprised that the SERP rankings fell off the face of the earth. Anyone have any ideas how I can recover from this? I've waited a couple of months since the changes took effect just in case Google was taking time to check it out.

    Read the article

  • CSS padding displays in FF and Chrome, but not in IE 8? [migrated]

    - by bullitt five
    I'm updating the CSS on a page design, trying to put borders around my images, with 7px of padding between the image and the border. It seems to be working fine in Firefox and Chrome, but IE displays the border directly against the image, with no padding. Any suggestions? CSS code: img.right { float: right; margin: 0px; border: 1px solid #999; padding: 7px; margin-left: 10px; margin-bottom: 5px; } HTML: <img src="images/homepage_challengecoin.jpg" class="right">

    Read the article

  • List of SEO tools [on hold]

    - by Felix
    I'd like to crowdsource a list of SEO platforms/tools. There is an abundance of options out there... Analytics SEO Blueprint BrightEdge Conductor Ginzametrics gShift Labs RankAbove Raven Internet Marketing Tools Rio SEO Searchmetrics seoClarity SEOlytics SyCara For each tool suggestion, please provide a brief overview of what the tool is used for and what differentiates it from its competitors.

    Read the article

  • Facebook likes reset after moving to HTTPS

    - by aarondicks
    I've got a question regarding the Facebook like button. We worked on a piece recently that embeds a number of social share buttons (please see the source code below). When the piece was released, it was on HTTP, and received over 2k likes (the URL 'slug' hasn't changed at all). The site was recently migrated to permanent-on HTTPS, and the like data has been reset, and we've been left with 50 new, recent likes. If you see in the source code, the URL is set explicitly to like the HTTP version, which I believe to be correct. Can anyone help me work out what's happened here? Here's the HTML bit of the like button: <div class="fb-like" data-href="http://www.harveywatersofteners.co.uk/history-interior-design" data-layout="box_count" data-action="like" data-show-faces="false" data-share="false"></div>

    Read the article

  • Trying to retrace our SEO domain redirect strategy

    - by dans
    An SEO built a copy of my company's e-commerce site on another domain that contained our product's keywords in the name (i.e. as if Levi's built a duplicate site on bluejeans.com)...and then they referenced a lot of the images on the actual website from the other domain (as if Levis.com had images on it referenced like: img src="http://www.bluejeans.com/jeans-front.jpg"), but when you tried to reach the site by typing the name into the browser you would be redirected to the regular website, so the site wasn't really used for any purpose except I guess SEO. Since I didn't think this was doing anything GOOD for us at the time, I deleted the duplicate site and let the hosting on it expire, only to watch our search engine position rankings fall dramatically. Any ideas as to what was going on there? I want to get it back to understand its impact, but I don't know how it was set up. I contacted our host and they have no idea how it was set up. I suspect there was some sort of redirect in play, or something?

    Read the article

  • Repurposing CSS Class Selectors

    - by limefartlek
    I don't know what this technique is called, I've only seen it used. It's a way to repurpose the same selectors with CSS. For example if I create h1 { font-size:18px; color:#FFFFFF; font-family:Arial, Helvetica;margin:0; padding:0; } h2 { font-size:18px; color:#000000; font-family:Arial, Helvetica; font-weight:normal;margin:0; padding:0; } I can repurpose the h selectors with something like .whatever h1 { color: #000; font: 2.0em arial, sans-serif; background-color: #fff3ea; margin: 50px 0px 0px 50px; } .whatever h2 { color: #000; font: 1.7em bold arial, sans-serif; background-color: #fff3ea; margin: 25px 0px 25px 75px; } If h1 and h2 appear inside of a div called whatever, then they will assume those properties. You can do this with ID tags and class tags but I can't for the life of me remember how this is done. Any thoughts?

    Read the article

< Previous Page | 239 240 241 242 243 244 245 246 247 248 249 250  | Next Page >