Search Results

Search found 28992 results on 1160 pages for 'content pages'.

Page 388/1160 | < Previous Page | 384 385 386 387 388 389 390 391 392 393 394 395  | Next Page >

  • Artists and music - Need Help Deciding on a CMS

    - by infty
    A friend has asked me to build a site with the following options: staff members must be able to add new music and artists to the page a gallery must be provided - it is also good if each artist has the ability to have his/her own smaller gallery users must be able to vote for artists users must be able to alter in discussions (forums or comments sections) staff members must be able to blog staff members must be able to write articles I did a small project where i actually implemented all of these features, but I want to use an existing content management system for all of these features so that future developers can, hopefully, more easy extend the website. And also, so that I don't have to provide too much documentation. I have never developed a website using an external CMS like Drupal or Wordpress and after seeing hours of tutorial videos of both systems, I still can't make up my mind on whether i should : a) use Drupal 7 b) use Wordpress 3 c) create my own cms I can imagine that staff members would also want to create content using iPhone or android based mobile devices, but this is not a required feature. Can someone, with experience, please tell me about their experiences with larger projects like this? The site will have approximately 400 000 - 500 000 visitors (not daily visitors, based on numbers from last year in a period of 4 months)

    Read the article

  • Is Microsoft Bing Trying To Kill Open Office?

    <b>Katonda:</b> "Microsoft Bing has many flaws, but this one seems to be the most outrageous one. If you try to search for OpenOffice on Bing, it will not show you the actual OpenOffice.org website but will show pages from random websites like OpenOffice.com or other non-OpenOffice.org websites."

    Read the article

  • Becoming a Fedora contributor

    <b>Ian's Thoughts:</b> "Becoming a Fedora contributor was easier than I expected.... I could hardly believe my eyes, these pages were laying out a step-by-step process that users could follow in order to become a member of a certain Team. "

    Read the article

  • 7 Ways to Get Ranked in Google Within 24 Hours

    In order for a website to receive more targeted visits, it needs to get indexed by major search engines like Google. But if you don't know the right strategy, it can take weeks before search engine spiders crawl into your pages. Listed below are proven techniques on how you can get indexed in less than 24 hours.

    Read the article

  • What Are Search Engine Crawlers?

    People don't know how they get relevant results for their search queries on a search engine. Most of them believe that these websites were submitted to the search engine. Few others think that there is some software tool that is searching for the relevant websites. Robots and spiders are the software tools that keep on searching the web to find new pages.

    Read the article

  • Nginx config - serving index.html not working

    - by Bill
    I can't figure out how to redirect / to index.html. I've gone through the threads on serverfault and I think I've tried every suggestion including: rewrite statements within location / index index.html at the server level, within location / and within static content moving node.js proxy statements to location ~ /i instead of within location / Obviously something is wrong somewhere else in my configuration. Here is my nginx.conf: worker_processes 1; pid /home/logs/nginx.pid; events { worker_connections 1024; } http { include mime.types; default_type application/octet-stream; sendfile on; keepalive_timeout 65; error_log /home/logs/error.log; access_log /home/logs/access.log combined; include sites-enabled/*; } and my server config located in sites-enabled server { root /home/www/public; listen 80; server_name localhost; # proxy request to node location / { index index.html index.htm; proxy_set_header Host $http_host; proxy_set_header X-Real-IP $remote_addr; proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; proxy_set_header X-NginX-Proxy true; proxy_pass http://127.0.0.1:3010; proxy_redirect off; break; } # static content location ~ \.(?:ico|jpe?g|jpeg|gif|css|png|js|swf|xml|woff|eot|svg|ttf|html)$ { access_log off; add_header Pragma public; add_header Cache-Control public; expires 30d; } gzip on; gzip_vary on; gzip_http_version 1.0; gzip_comp_level 2; gzip_proxied any; gzip_min_length 1000; gzip_disable "msie6"; gzip_types text/plain text/css application/json application/x-javascript text/xml application/xml application/xml+rss text/javascript; } Everything else is working just fine. Requests get proxied to node correctly and static content is served correctly. I just need to be able to forward requests made to / to /index.html.

    Read the article

  • How to Check How Many Inbound Links Your Competitor Has

    Inbound links are often referred to as off page search engine optimization (SEO) and they are really important to get a high ranking in the search engine results. To rank and stay higher than your competitor, you will need a way to find out how many back-links they have built to their web pages. In this article you will get a quick and easy way on how to check how many inbound links your competitor has.

    Read the article

  • Why Doing SEO "The Right" Way Works

    Online searching has been found to be the third most frequent activity for people who log on to the net. This is why it is important to stay on top of the search engine result pages, as this will help the users to find the website, thus increasing qualified traffic.

    Read the article

  • How to give specific url using htaccess? [on hold]

    - by Dash
    I am a web developer using codeigniter.I want to give a specific url to certain pages on my website.Is it possible using htaccess?I visited following sites but couldn't find anything such there: Bluehost Tutplus and some others too. What i really wanna do is when the admin is logged in the link should be http://localhost/admin-ci/index.php/admin/index.php/dashboard and if user logs in then the link should be http://localhost/admin-ci/user/index.php/dashboard.Will htaccess be able to do this?

    Read the article

  • MySQL performance over a (local) network much slower than I would expect

    - by user15241
    MySQL queries in my production environment are taking much longer than I would expect them too. The site in question is a fairly large Drupal site, with many modules installed. The webserver (Nginx) and database server (mysql) are hosted on separated machines, connected by a 100mbps LAN connection (hosted by Rackspace). I have the exact same site running on my laptop for development. Obviously, on my laptop, the webserver and database server are on the same box. Here are the results of my database query times: Production: Executed 291 queries in 320.33 milliseconds. (homepage) Executed 517 queries in 999.81 milliseconds. (content page) Development: Executed 316 queries in 46.28 milliseconds. (homepage) Executed 586 queries in 79.09 milliseconds. (content page) As can clearly be seen from these results, the time involved with querying the MySQL database is much shorter on my laptop, where the MySQL server is running on the same database as the web server. Why is this?! One factor must be the network latency. On average, a round trip from from the webserver to the database server takes 0.16ms (shown by ping). That must be added to every singe MySQL query. So, taking the content page example above, where there are 517 queries executed. Network latency alone will add 82ms to the total query time. However, that doesn't account for the difference I am seeing (79ms on my laptop vs 999ms on the production boxes). What other factors should I be looking at? I had thought about upgrading the NIC to a gigabit connection, but clearly there is something else involved. I have run the MySQL performance tuning script from http://www.day32.com/MySQL/ and it tells me that my database server is configured well (better than my laptop apparently). The only problem reported is "Of 4394 temp tables, 48% were created on disk". This is true in both environments and in the production environment I have even tried increasing max_heap_table_size and Current tmp_table_size to 1GB, with no change (I think this is because I have some BLOB and TEXT columns).

    Read the article

  • Search Engine Optimization Basics - Know Your Keywords

    I laugh to myself when I hear everyone in SEO talking about how the most important thing to improve your website's position in Google is backlinks. Now don't get me wrong, I definitely agree that building backlinks to your pages is important, but there is something that I consider much more important - and that is keyword research.

    Read the article

  • Benefits of PSD to HTML Service? For Whom and How

    With the advent of Internet and e-industry, most of the companies create website and hire web development professionals. And it is also true that the process of converting a design into web pages is ... [Author: Manish Rawat - Web Design and Development - June 13, 2010]

    Read the article

  • Create an RSS feed for any webpage, monitor changes and send results by email?

    - by David
    Do you know a Windows 7 software (not an online service) than can create an RSS feed for any webpage ? Here is what I am looking for : I manually select what is a title, what is a link and if needed what is the content that I would like to keep (article content + their related images if they exist). The software create a RSS feed from that. Then it monitors the feed every x hours (or x days at a specified time). If updates are found then send the results to my email address. (It is a little like a combination of the Firefox addon: Autopager (WYSIWYG selection which use xpath) and WebSite-Watcher (RSS software monitoring).

    Read the article

  • Google projette d'instaurer des bannières publicitaires sur son moteur de recherche, en 2005 l'entreprise avait pourtant promis ne jamais le faire

    Google projette d'instaurer des bannières publicitaires sur son moteur de recherche, en 2005 l'entreprise avait pourtant promis ne jamais le faire En 2005,Marissa Mayer qui était alors vice-présidente sur les produits de recherche et l'expérience utilisateur chez Google, a affirmé dans sur un blog officiel qu'il n'y aura « pas de bannières de publicité sur la page d'accueil de Google ou sur les pages de résultats de recherche. Il n'y aura pas de publicités folles, criardes, graphiques qui voleront...

    Read the article

  • Choosing the Right Company For Your SEO Needs

    One important factor that determines or contributes to the success of a website is through SEO services. In a world where competition is rife and where one has to keep on promoting their product constantly, advertising has emerged as the main avenue for making your products known to many potential customers. Some years back, businesses and individuals advertised their services and products in the yellow pages or newspapers.

    Read the article

  • Apache Bench length failures

    - by Laurens
    I am running Apache Bench against a Ruby on Rails XML-RPC web service that is running on Passenger via mod_passenger. All is fine when I run 1000 requests without concurrency. Bench indicates that all requests successfully complete with no failures. When I run Bench again with a concurrency level of 2, however, requests start to fail due to content length. I am seeing failures rates of 70-80% when using concurrency. This should not happen. The requests I am sending to the web service should always results in the same response. I have used cURL to verify that this is in fact the case. My Rails log is not showing any errors as well so I am curious to see what content Bench actually received and interpreted as a failure. Is there any way to print these failures?

    Read the article

  • Google analytics campaign advice

    - by Drewsdesign
    I am buying traffic from a broker not one source and sending to various landing pages. I would like to know the best way to structure a campaign so I can find which referrering site/url is performing the best (time on site, bounce etc) Should the utm_campaign be the 'brokername' and the utm_source be the 'landingpagename' or should this be the other way around? Also what would be the best way to create a custom report to show all the referrers metrics by each landing page ? Thank guys really appreciate any help on this.

    Read the article

  • Quantcast Media Planner

    Have you ever wondered what type and how many people are visiting your website? If you answered "yes" to the previous question, you will be pleased to know that there are online tools available that allow people to analyse the relevance and effectiveness of their web pages.

    Read the article

  • How to reset a List c# and XNA [on hold]

    - by P3erfect
    I need to do a "retry" option when the player finishes the game.For doing this I thought to reset the lists of Monsters and other objects that moved at the first playing or which have been "killed".for example I have a list like that: //the enemy1 class is already done // in Game1 I declare it List<enemy1> enem1 = new List<enemy1>(); //Initialize method List<enemy1> enem1 = new List<enemy1>(); //LoadContent foreach (enemy1 enemy in enem1) { enemy.Load(Content); } enem1.Add(new enemy1(Content.Load<Texture2D>("enemy"), new Vector2(5900, 12600))); //Update foreach (enemy1 enemy in enem1) { enemy.Update(gameTime); } //after being shoted the enemies disappear and i remove them //if the monsters are shoted the bool "visible" goes from false to true for (int i = enem1.Count - 1; i >= 0; --i) { if (enem1[i].visible == true) enem1.RemoveAt(i); } //Draw foreach (enemy1 enemy in enem1) { if(enemy.visble==false) { enemy.Draw(spriteBatch, gameTime); } } //So my problem is to restart the game. I did this in Update method if(lost==false) { //update all the things... } if(lost==true)//this is if I die { //here I have to put the code that restore the list //I tried: foreach (enemy1 enemy in enem1) { enemy.visible=false; } player.life=3;//initializing the player,points,time player.position=initialPosition; points=0; time=0; }//the player works.. } } they should be drawn again but if I removed them they won't be drawn anymore.If I don't remove them ,instead, the enemies are in different places (because they follow me). Any suggestions to restore or reinitialize the list??

    Read the article

  • How to make my website loads fast [duplicate]

    - by Garva Sharma
    This question already has an answer here: Ideas to improve website loading speed? 1 answer this is my website nxttech.org and it loads really slow so please review it and tell me what can i do with it to make it faster. And my websites some pages loads fast while some loads really slow so what is this does this is normal for a website or it is issue with my hosting service.

    Read the article

  • Alias one set of Subdomains to another

    - by Schneems
    I have to domains that I want to effectively mirror one another on select subdomains. Let's say they are pirates.com and ninjas.com and i want ninjas to mirror the content on pirates. When I visit foo.ninjas.com I want to see the content on foo.pirates.com. Due to app restrictions I need to do this in DNS for an undefined number of subdomains. I was under the impression i could do this with a CNAME, but it appears that setting a CNAME for * subdomain of the ninjas.com domain to point to pirates.com will make any subdomain of ninjas.com point at pirates.com instead of the associated subdomain. I.e. foo.ninjas.com would reference pirates.com instead of foo.pirates.com. Is there a way to do this using DNS? Am I missing something basic?

    Read the article

< Previous Page | 384 385 386 387 388 389 390 391 392 393 394 395  | Next Page >