Search Results

Search found 96383 results on 3856 pages for 'code pro'.

Page 627/3856 | < Previous Page | 623 624 625 626 627 628 629 630 631 632 633 634  | Next Page >

  • Hosting a website with Windows Azure

    - by Rev
    I may be completely misunderstanding what Azure is but is it possible for me to host a basic website on Windows Azure? I have a site that I've built in HTML and CSS that I'd like to upload to Azure but I can't figure out any way to do this. The site claims I can use it for web hosting, but if I can't FTP then I'm not sure how to do this. Is there a simple tutorial somewhere? I couldn't find anything close to what I'm looking for through searching.

    Read the article

  • What to use to make voice chat (and some more) on a web?

    - by Tunococ
    I am trying to make available on my website a voice chat for a small group of people that allows some other means to interact such as text messaging, photo sharing, file sharing, simple drawing and silly games. In other words, something similar to older MSN Messenger, but on the web. Any ideas on what to use? To clarify, I am looking for suggestions on languages and libraries to use. I want to be able to fully customize it as much as possible because I might want to add other (somewhat interesting) functions later. Low-level programming is fine if required, but platform dependency isn't that much preferred.

    Read the article

  • What is this thing called?!?! (bottom of screen popup)

    - by NRGdallas
    I am looking for the name and some possible jquery libraries etc for the standard bottom of screen popup bar. Its like a little bar that pops out on the bottom of the screen after X seconds or however really - generally slides up, about 50px or so high, and usually the length of the main container. used for some form of coupon advertisement or various promotion text etc. What is the proper term for this item, and would there be any good references to best-use guidelines?

    Read the article

  • fault tolerant uploading tool

    - by andersonbd1
    I'm setting up a wordpress site for a friend who has some somewhat large audio files (150M)... He's on a bad connection and I'd guess it'll take him a while to upload those files with the normal wordpress way. I'm looking for a tool that I can install on the server that allows uploads and is also fault tolerant... for example if you loses his connection, or power, or whatever it'll pick up where it left off. I realize web technologies probably don't do that, but perhaps flash or something? Any ideas?

    Read the article

  • Since Google doesn't use nofollow anymore what is smart alternative?

    - by SharkTheDark
    Since from 2009. Google count nofollow links also as outgoing link, what are alternative to stop Google count outside links from my page? If I make links appear on my page source like this: <span hrefs="http://link" rel="nofollow" link="true">Link Name</span> and then in JavaScript replace span with a tag and replace hrefs with href for every span tag that has link="true". Will this help?

    Read the article

  • Is it possible to force a tab/window to take focus without an alert()? [migrated]

    - by Chris Brandt
    I've been working to solve some form / session timeout issues on our site. I'd like to get the users attention BEFORE this happens so that they don't lose data that they've entered into long forms. I'm using jquery's .get to confirm with my server when the users session is set to timeout, and warn them 5 minutes before then. I've found that if I pop an alert(), it will bring the focus back to that tab / page -- but this seems 'heavy handed'. Is there a better way to do this?

    Read the article

  • .htaccess redirect question

    - by user473056
    Hi, I'm trying to set up my .htaccess file to take the displayed link and route it to the destination link as below Displayed Link http://www.my-website.com/click-4559226-10388358?url=https%3A%2F%2Fdestination-website2.com%2FItem.php%3Fid%3D44350396%26sld%3DA6D7A632-821E-4b78-ACD0-147658B77BD6 Destination Link http://www.destination-website.com/click-4559226-10388358?url=https%3A%2F%2Fdestination-website2.com%2FItem.php%3Fid%3D44350396%26sld%3DA6D7A632-821E-4b78-ACD0-147658B77BD6 Effectively, all that changes is the first url (http://www.my-website.com) everything after that is the same. Is this possible and could someone briefly explain how I would go about it? * Just to be clear, I dont want to redirect everything from my-webiste.com. Just links that start http://www.my-website.com/click-4559226-10388358

    Read the article

  • Is the php method md5() secure? Can it be used for passwords? [migrated]

    - by awiebe
    So executing a php script causes the form values to be sent to the server, and then they are processed. If you want to store a password in your db than you want it to be a cryptographic hash(so your client side is secure, can you generate an md5 using php securely( without submitting the user:password pair in the clear), or is there an alternative standard method of doing this, without having the unecrypted pasword leaving the clients machine? Sorry if this is a stupid question I'm kind of new at this. I think this can be done somehow using https, and on that note if a site's login page does not use https, does that mean that while the databse storage is secure, the transportation is not?

    Read the article

  • Analytics - Total events divided by number of unique pages?

    - by GeekyAndUnique
    I am using Google Analytics events to track keywords on my articles - not necessarily the best system I know but there are too many for variables I can't easily change it right now - and I would like to be able to see how popular each keyword is by dividing the number of page views with a keyword by the number of unique pages. Is there a/what is the best way of doing this? EDIT FOR CLARITY I currently have a system set up where every time somebody loads an article an event is fired for each of the tags/keywords used, with the keyword being the label. I can currently view my view count for each of the keywords by looking at the total events for each label, however I would like to be able to see which keywords are the most popular by dividing the number of times the event has been fired by the the number of different pages it has been fired from.

    Read the article

  • CentOS drive mapping? [on hold]

    - by DroidOS
    This is the first time I am posting on this particular StackExchange forum and I hope that I am using the right one for the present question. Briefly, this is what I need to do I am running a web service where users can, amongst other things, upload and store files on the server. What I want to do is to hive off user file storage to a different location so my server (CentOS 64 bit) can concentrate on what it does best - server side scripting and database management. As things stand all user files go into subdirectories in a folder called stash that lies above DOC_ROOT. What I would like to do is Transparently detect all attempts to read/write to stash/sub_folder and get/set file data on a remote server - ideally the latter would be one which replicates files like a CDN so they can be delivered from the closest/fastest location based on where the user's location. Even nicer would be if for all read accesses I could provide a URL that allows the user's browser to fetch the relevant file directly without having to funnel them via my server. I am a relative newbie when it comes to this sort of thing so I hope that I have phrased this question adequately well. From the little searching I done I gathered that WebDAV can be used to map drives to a different location on the web so perhaps that is a starting point. But if that will work I need to Establish how to get WebDAV up and running on my CentOS 64 bit server. Ideally, identify a service that allows this kind of file storage and provides an API I can use in my own scripting. I'd much appreciate any help with this.

    Read the article

  • AGLS Metadata - Is it widely adopted?

    - by Brandrally
    Recently, I have seen in a couple sites around Australia's meta data AGLS tags. <meta name="AGLS.Audience" scheme="agls-audience" content="All"/> <meta name="DC.Publisher" scheme="AglsAgent" content="Hyundai"/> I have never seen this kind of mark-up before and discovered: http://www.agls.gov.au/ Just wondering whether there is a big community / support out there for the adopting these tags? Any thoughts would be great.

    Read the article

  • Transferring users and search engines to a new domain

    - by eftpotrm
    I've been asked to take over the maintnance of an existing site that's being reworked. At present it's serving localised content for several languages, but via a fairly unhelpful mechanism that means essentially search engines only have it indexed in English and any deep links will de facto appear in English as well. So, new localised sites are being built under separate domains - not just for this, there's other benefits. What we're then looking to do is to redirect users correctly to the new site, where appropriate. For humans this isn't a problem. We can send them through a gateway page on their first site visit, grab their language preference and put it in a cookie, then redirect them to the new localised content as soon as it's available. For search engines, this isn't so good... In principle I'm happy to simply bypass the gateway page and redirect known spiders to the new site, but this means we're serving radically different content (different URL even!) to human and robot users. Won't this therefore be regarded as cloaking and cause us grief? Anyone know a better way to handle this?

    Read the article

  • How should I design my website to allow posterity to edit?

    - by SSumner
    I'm building a website for a student organization I am involved in at my college. Most of the site will be static - i.e. won't change from year-to-year, but certain pieces will. I am high-tech, but most of the others aren't, and I am graduating in the spring. So how should I go about building the website so as to allow those that take over in subsequent years to edit information? Examples: Events: I already plan on using a Google calendar for this Officers: There will be profiles/pictures for all the officers on the web page Connections: Partnerships with other organizations that we have currently, but may not in future, or may add more in future Should I use some form of CMS (Content Management System)? If so, how restrictive are they (e.g. Drupal) to what you can build and then how easy is it to edit. What other ways could I make a very nice-looking website but allow certain pieces to be edited later?

    Read the article

  • How to make Google show my site in search result like the following image? [closed]

    - by Samik Chattopadhyay
    Possible Duplicate: What are the most important things I need to do to encourage Google Sitelinks? Currently Google is displaying my site (http://layzend.info) like this in search result, only the link and meta description without any internal page links - But I want to be the search result like the following where the internal links are also displayed - How is it possible? Please help me to make my site more SEO friendly.

    Read the article

  • What is the ideal file size for a web page? [closed]

    - by Rob
    Possible Duplicate: Is there a maximum size that web pages should be kept under? What is the ideal file size for a web page? Specifically when it comes to image sizes, what's the total file size for a webpage which includes several images. I tend to compress images down as much as possible before it starts to visually lose quality. We run several CMS website's and the clients tend to ask this question a lot! I'd love to hear another view on it.

    Read the article

  • Will a search engine lower the rank of my page if i have a hidden iframes?

    - by Skurpi
    As a praxis, all external content on our site is put in iframes to lower the risks of any external parties injecting stuff to our users. We also do it to make sure our content shows up before banners, to make the site feel quicker. We now have an external script running which we want to put in an iframe, but it does not have any visible content to go with it so I want to put css "visibility: hidden;" on the iframe. I read in a forum somewhere that search engines will lower the rank of a page, or even drop the page, if a iframe has "the minimal size of 1x1px". Will a search engine lower the rank of my page if I have a hidden (or 1px big) iframe?

    Read the article

  • SEO and search result changes when switching to SSL on Joomla site?

    - by jeffery_the_wind
    I am thinking about purchasing an SSL certificate for a website. The most noticeable difference for the user would be the http now becomes https and there is that lock icon in most browsers. Will there be any adverse affects on the website's current SEO or recognition by search engines when I make the switch? Also this is a Joomla site, which has an option in the settings to use SSL. It is supposed to make it easier but not sure if it takes care of everything. Thanks!

    Read the article

  • wamp installed can't find localhost

    - by Tac
    I installed a WAMP package on a Windows 7 machine and everything ( Apache, php, mySql, phpMyAdmin) appear to have installed correctly. However when I try to access localhost or phpMyAdmin via the browser, I get "Server Not Found". I've tried using localhost, 127.0.0.1 in the browser. I've checked the httpd file it says "Listen 80". C:\Windows\System32\drivers\etc\hosts says "127.0.0.1 localhost" Apache log file doen't show any errors. Suggestions appreciated.

    Read the article

  • CMS/platform/portal for personnel placement/recruiting [closed]

    - by mezzanine
    Possible Duplicate: Which Content Management System (CMS)/Wiki should I use? I am interested in building a platform where I can match specialists/experts in a certain profession with potential customers/clients. Therefore, the professionals should be able to register with information about them and their area of expertise, and customers should be able to search for them and contact them via my website. Bonus would be some kind of feedback or rating functionality. Examples for other areas are http://www.lawyers.com/ for lawyers and http://sortfolio.com/ for web designers. Is there some kind of platform which already provides this functionality and could be customized?

    Read the article

  • Google indexing pages with #! although we don't have any

    - by Benjamin Gruenbaum
    Our company has developed a Single Page Application using AngularJS and its routing. Google indexed our site decently with JavaScript but it did not index some pages very well so we have developed an HTML only version. We have followed the Ajax Crawling Specification posted here and have a <meta name='fragment' content='!'> tag and canonical urls. We expect http://www.example.com/foo/bar to be fetched from http://www.example.com/?_escaped_fragment_=/foo/bar. However, we have found out that when we rolled the AJAX specification we now have all pages indexed twice, once with the JavaScript version as http://www.example.com/foo/bar and once with the new version as http://www.example.com/#!/foo/bar. This is harmful to us since it's duplicate content and also mis-representing out site. I have tried looking for similar questions here and in the Google product forum but could not come up with anything.

    Read the article

  • Which is the best way to catch an expiring domain name? [closed]

    - by newspeak
    I know a similar question has been asked, but I really don't know what to do. There is this .com domain which is currently on redemption period and should likely be available again within a month. I was wondering which is the best way to get it at a reasonable price. I don't think it's a highly valuable domain, it shows to have very bad ranking and has 0 exact same searches according to adwords. Why it is valuable to me is very simple: I have a project responding to this name. I already own the .net domain and would love to have the .com. I discovered the domain was going to be available thanks to an email I received by a backorder site. I did some research and these guys have a bad reputation on the web. I did further research and found that more reputable (at least in theory) companies should be the likes of snapnames, pool, namejet, godaddy, etc. I am a bit suspicious using these drop cathing services: What if they shill bids? What if they make it go into auction even if I'm the only person interested? What if I raise attention and interest to the domain by backordering? I just would rather wait for it to be deleted and available again to register it manually. It is really not an interesting domain name, and I don't think anyone would care to have it. But what if the domain is already being watched by the domain industry sharks? I did a whois research and my desired domain nameserves point to domcollect.com, which appears to be an auction site. What if I decide to wait for manual registration and I miss the chance to get it? I'm willing to spend the 60/70$ fees these sites require, but not really more than that. Suggestions? Thank you very much. I'm a bit confused and undecided.

    Read the article

  • Should I design and then look for a CMS or vice versa? [closed]

    - by Livingston Storm
    I am currently designing an e-commerce site, and unfortunately my PHP is garbage at the moment so open source CMS's are out of the question. I am debating between Joomla and Big Commerce and as the title states I am unsure of whether to build first or try the CMS first to see what limitations I will face. I couldn't find any previous questions on this site, forgive me if this is a stupid/commonly asked question. Thanks for any feedback

    Read the article

  • pages still show up in google search even after disallowed in robots.txt [duplicate]

    - by Jota Onasys
    This question already has an answer here: With Robots.txt disallow all, why was my site still getting traffic? 5 answers Why is it that some pages still show up in google search even though disallowed in robots.txt? Is the best solution here to remove the Disallow from Robots.txt and just add noindex, nofollow meta tag to those pages you want blocked? Or should I submit a request to Google directly to remove those pages?

    Read the article

< Previous Page | 623 624 625 626 627 628 629 630 631 632 633 634  | Next Page >