Search Results

Search found 21219 results on 849 pages for 'content expiration'.

Page 433/849 | < Previous Page | 429 430 431 432 433 434 435 436 437 438 439 440  | Next Page >

  • How do I grep in a variable?

    - by Ashfame
    How to do a grep in a variable? I have stored the wget output in a variable and I need to extract out some strings from it. Like the content of the variable is upgrade http://wordpress.org/download/ http://wordpress.org/wordpress-3.0.5.zip 3.0.5 en_US 4.3 4.1.2 I need to check if the string contains the word upgrade, so I can do a simple grep and then check the exit status of it by $? and proceed. But how can I get the value 3.0.5 which is actually the fourth word? And how to actually grep in a variable?

    Read the article

  • What is the advantage to hosting static resources on a separate domain?

    - by Michael Ekstrand
    I notice a lot of sites host their resources on a separate domain from the main site, e.g. StackExchange using sstatic.net, Barnes & Noble using imagesbn.com, etc. I understand that there are benefits to putting your static resources on a separate host, possibly with an efficient static-file web server like nginx, freeing up the main server to focus on serving dynamic content. Similarly, outsourcing to a shared CDN like cloudfront Akamai is logical. What is the benefit to using a separate domain otherwise, though? Why sstatic.net instead of static.stackexchange.com? Update: Several answers miss the core question. I understand that there is benefit to splitting between multiple hosts — parallel downloads, slimmer web server, etc. But what is more elusive is why multiple domains. Why sstatic.net rather than static.stackexchange.com as the host for shared resources? So far, only one answer has addressed that.

    Read the article

  • Oracle University Nuovi corsi (Week 42)

    - by rituchhibber
    Oracle University ha recentemente rilasciato i seguenti nuovi corsi in inglese: Database Oracle Enterprise Manager Cloud Control 12c: Install & Upgrade (Training On Demand) MySQL Performance Tuning (Training On Demand) Fusion Middleware Oracle GoldenGate 11g Fundamentals for Oracle (4 days) Oracle WebCenter Content 11g: Site Studio Essentials (5 days) Oracle WebCenter Portal 11g: Build Portals with Spaces (3 days) Business Intelligence Oracle BI 11g R1: Create Analyses and Dashboards (4 days) SOA & BPM SOA Adoption and Architecture Fundamentals (3 Days) eBusiness Suite R12 Oracle Using and Maintaining Approvals Management - Self-Study Course R12 Oracle HRMS Advanced Benefits Fundamentals - Self-Study Course WebLogic Oracle WebLogic Server 11g: Monitor and Tune Performance (Training On Demand) Financial Oracle Project Financial Planning 11.1.2: Create Projects ( 3 days) Tuxedo Oracle Tuxedo 12c: Application Administration (5 days) Java Java SE 7: The Platform Evolves - Self-Study Course Primevera Primavera Client/Server Partner Trainer Course - Self-Study Course Primavera Progress Reporter 8.2 - Self-Study Course Per ulteriori informazioni e per conoscere le date dei corsi, contattate il vostro Oracle University team locale.

    Read the article

  • Website Stopped Showing From Google Search Results Sunddenly

    - by Aman Virk
    I have a design and development blog http://www.thetutlage.com (1.5 years old), which was doing really well in Google search as I was getting over 70% of my traffic from Google. Now suddenly from last two days it reduced the amount of traffic from 70% to 20% and also when I am trying to search for the exact posts that I can created even after appending my website name to it does not show any results for that. Sample Search Text: JQuery Game Programming Creating A Ping Pong Game Part 1 I have post with exact same title and it does not show it on Google search anywhere. I am totally shocked, I write my own unique content and follow Google guide lines like bible. Also there is no message under my webmasters account stating any problem or error.

    Read the article

  • SPARC Servers at Oracle OpenWorld

    - by B.Koch
    There is plenty to learn about the SPARC servers at the Oracle OpenWorld. The SPARC server sessions offer depth and breadth in content selection to satisfy everyone's need from the one who is technically oriented to the one who would like to understand more about the business value of SPARC technology. And, there is always more. Keynotes, Oracle innovations and many product demonstrations are only a few of many other opportunities to interact with the product experts and executives to establish greater insight to the Oracle SPARC technology. The Oracle SPARC Servers Oracle's SPARC servers running Oracle Solaris are ideal for mission-critical applications that require high performance, best-in-class availability, and unmatched scalability on all application tiers. With a robust roadmap, Oracle assures the highest levels of investment protection through 100% SPARC/Solaris binary compatibility, proven by hundreds of thousands of deployments over more than 20 year. 

    Read the article

  • Job Search Engine Url Structure Issue [closed]

    - by Justin
    Possible Duplicate: What is the best stucture of SEO friendly URL? I am working on a job board, and i'm trying to figure out a good design for URL structure. Some things that I have found through research: 100 - 150 Chars long is ideal 3-5 words in your url, according to Matt Cutts Use .htaccess to force clean urls Do not duplicate data (important) Clean and precise, describing the content Use hyphens On the homepage, I try to detect the users location based on IP, but this isn't always accurate, and not always reliable. So until they put in their city/location, I can't always use this structure but this is potentially work-able. For Searching, a form post to a results page: domain.com/jobs/[city]/[search] ie: domain.com/jobs/toronto/sales manager/ OR domain.com/search/jobs/toronto/sales manager/ or do I remove the word JOBS and just use Search. I trying to keep good search terms in the URL, but also keep it clean and concise. Can someone give me some feedback and thoughts to 'why'...

    Read the article

  • Should I have link rel=next & prev on URLs which have query variables?

    - by user21100
    For example, I have link rel prev & next set up on these pages of products: site.com?page=2 site.com?page=3 (this is my preferred structure by the way and I'm trying to get all the ugly URLs which are littered with query variables deindexed as they are causing duplicate content). So the above URLs are fine but once a filter to narrow product results is selected, like "price", the URL shows like this: site.com?price[1000-1499]=on site.com?page=2&price[1000-1499]=on As of right now, I am having the link rel prev & next dynamically added to the header of these pages but since I am working on getting these query variable URLs pages deindexed, I am wondering if I should get rid of it on these pages? Any thoughts?

    Read the article

  • Grep in a variable

    - by Ashfame
    How to do a grep in a variable? I have stored the wget output in a variable and I need to extract out some strings from it. Like the content of the variable is upgrade http://wordpress.org/download/ http://wordpress.org/wordpress-3.0.5.zip 3.0.5 en_US 4.3 4.1.2 I need to check if the string contains the word upgrade, so I can do a simple grep and then check the exit status of it by $? and proceed. But how can I get the value 3.0.5 which is actually the fourth word? And how to actually grep in a variable?

    Read the article

  • Why Ubuntu is not booting anymore from USB live?

    - by xRobot
    I have just purchased a brand new laptop Samsung 300e5c with Windows 7. So I have reduce the windows partition and I have installed ubuntu 12.04 from my USB pen drive. Then I have tried to boot again from USB pen drive but it doesn't work anymore. I have tryed to boot on others laptop from my USB pen drive and it does work perfectly. but on my laptop, NO :(. I have set all USB drive as first boot device in BIOS. I have tryed on each USB port of my laptop but it doesn't work. Only the first time it did work when I installed ubuntu and now it doesn't work anymore. It's very strange I can see the content of the usb pen drive on my laptop and on others laptop without problems. why it doesn't work anymore ?

    Read the article

  • Is it safe to Block These URLs with Robots.txt?

    - by Edgar Quintero
    I have a website that has all URLs optimized and 301 redirected from nasty URLs to clean ones. However, everywhere throughout the site the unclean URLs are linked in menus, content, products, etc. Google currently has all clean URLs indexed, along with a few unclean URLs too. So the site still has linked everywhere the old URLs (ideally this wouldn't be the case but this is how it is ATM). I would like to block the unclean URLs with robots.txt. The question: If I block these unclean URLs with the robots.txt, when the entire website is linked with them (but they all redirect to the clean version), will this affect the indexing status at all?

    Read the article

  • So long Oracle

    - by Roger Brinkley
    This Friday (November 1, 2013) is my last day at Oracle. After Publishing almost 150 Java Spotlight Podcasts with over 800,000 downloads plus another 95 Mobile and Embedded Podcasts Being a founding member of Java.Net with the Desktop Community Starting and founding 3 separate Java.Net Communities. Visiting every continent save one and several cities around the world Traveling to more cities in Brazil than most Brazilians. Speaking at all major Java conferences and lots of Java User Groups Being the longest serving JavaOne Track Lead Meeting, working, and interviewing the best of the best in the Java community Creating JavaHelp the online help system for the Java Platform Most importantly having lots of fun (who else played hickory golf in 6 continents in one year) Its time for me to move on! I'll now be publishing content at http://www.rogerbrinkleyblog.info/ now onwards. Read my new blog to learn about my new adventures!

    Read the article

  • SEO With a Linkwheel

    You will be amazed to know that more than 68 % of internet users do not go farther than first page on search results. That means you have to be among the top page ranks to have any business online. If you are not among those listed on the first page of search results, you are missing out on most of the users searching for products, information and content on Google and other search engines. Linkwheel is a wonderful strategy for search engine optimization (SEO). If done well, your search engine rankings will skyrocket in a very short time.

    Read the article

  • How to use rel=canonical with Sitecore aliases?

    - by Mike G
    I have inherited a Sitecore architecture that is a mess from an SEO duplicate content POV. There are multiple aliases that have been created (and indexed by the search engines) for many of the 2nd tier pages of the site. Due to server issues, I am not able to 301 redirect these duped pages, so I would like to use the rel=canonical tag in an attempt to try and get Google/Bing to recognize the correct pages I would like to appear in the index. I have blocked the most extraneous duped pages with a robots.txt file, however, since Google/Bing have already spidered many of the duped pages, I need to keep them accessible to the spiders, BUT removed from the index. The catch is, since the duped pages are aliases (and don't really physically exist in Sitecore that I can find), I am not sure how to go about using rel=canonical - or if I even can in this situation..?

    Read the article

  • joomla sometimes messes up urls, probably cache involved

    - by Bakaburg
    Is a bit i'm having this problem and i really cannot get the hang of it... Every once in while my joomla site messes up links url and for example from something like this: http://www.sism.org/index.php?option=com_comprofiler&task=userslist&listid=4&Itemid=123 it becomes like this: http://www.sism.org/index.php/component/k2/administrator/components/com_dump/assets/css/images/stories/inrilievo/sism/htm/index.php?option=com_comprofiler&task=userslist&listid=4&Itemid=123 the new page has the right content but there are no css and other linked resources. Usually i solve the problem by deleting all the cache and turning it off and on again. Of course this is pretty annoying especially for my association. Does any one have any clue on this? Watching the URLs the components involved seems to be K2 and Jdump. Thanks

    Read the article

  • What is the current legal status of magnet links?

    - by Moonwalker
    Prelude: I develop a cloud service which could be described as dropbox meets torrents and as side effect it enables distribution of arbitrary content via magnet links. Certain amount of magnet links will be displayed on the main website (I will be able to remove them one-by-one or ban users but no more). I will not be able to avoid magnets without complete rework of overall project architecture and either way it will hurt overall performance badly, probably making service meaningless. So my question is, what should I do, to avoid legal problems if my site in a nutshell is just a collection of magnet links? Privacy achieved via end-user encryption, so there is almost no access restrictions on the website. And anyway will help me any? Will hosting in particular country help me?

    Read the article

  • How would I pursue a track in front-end web development?

    - by Koviko
    I've recently been put on heavy JavaScript projects and have become fond of the front-end world in comparison to the back-end. I have always been good at proper markup and CSS, and coupled with AJAX, pretty animations, and dynamically generated content, it's become a much more interesting and flashy world for me. I would like to be able to continue to hone my craft in the same way that I was able to become proficient at back-end development with PHP: getting paid to do it. How would I market myself as a front-end web developer with a strong interest in dynamic JavaScript-driven websites? Due to my strong background in back-end development, how would I find the companies that wouldn't waste my front-end skill set on simple HTML/CSS development? And as a bonus, how would I apply this to being a contractor/freelance developer rather than a salaried employee? While I like the idea of being able to remain a part of my creations, I also dislike the maintenance phase of projects.

    Read the article

  • Issue in Webscrapping in C# : Downloading and parsing zipped text files

    - by user64094
    I am writing an webscrapper, to do the download content from a website. Traversing to the website/URL, triggers the creation of a temporary URL. This new URL has a zipped text file. This zipped file is to be downloaded and parsed. I have written a scrapper in C# using WebClient and its function - DownloadFileAsync(). The zipped file is read from the designated location on a trapped DownloadFileCompleted event. My issue : The Windows 'Open/Save dialog is triggered". This requires user input and the automation is disrupted. Can you suggest a way to bypass the issue ? I am cool with rewriting the code using any alternate libraries. :) Thanks for reading,

    Read the article

  • Web App vs Portal Platform - convincing the customer

    - by shinynewbike
    We're evaluating a set of requirements for a customer who wants Liferay which mainly has AAA and Web CMS requirements, and allowing user to upload their own content. Also all inetgration is via web services. However there is no need for other features such as actual "portlets", i18n, mashups, skins, themes, tagging, social presence, no collaboration etc So we feel we can do this as a standard JEE web app and not use Liferay (or any other portal product) since these are overheads we dont need. The customer feels the Web CMS requirements + user upload justify the "portal" product. Can anyone help me with some points to convince the customer? Assuming our point of view is right.

    Read the article

  • Is there any way to discover the traffic of a site I don't control?

    - by George Bailey
    Given the following: The website does not call any external images or scripts, all the content is hosted on a server that is in our control. The website does not contain the meta tag, nor does it contain the html file that would authorize a Google Account access to Webmaster Tools. The access logs have not been provided to any 2nd or 3rd party. Is it possible for a 3rd party to get an idea of how many hits the site is getting, or are they limited to just seeing how high the site ranks? How could the 3rd party determine how well the site is doing under these restrictions? Is there a website for that that you know of?

    Read the article

  • Questions Anyone?

    - by Kevin Smith
    I've been working with WebCenter Content for almost 9 years now and have a ton of topics rolling around my head that I'm sure would make excellent blog post once I find the time to write them up. Does anyone have any questions they would like answered? Like why does WCC do ...? How does this feature work? I can't seem to get this working? Post your question to the comments and if it is something on my list of topics I had planned on creating a blog post about I will move it to the top of the list.

    Read the article

  • My VS 2010 and ASP.NET 4 Talks Online

    The past 7 years Ive done an annual all day event in Arizona organized by the most excellent Scott Cate (who always does a phenomenal job organizing the event and making it a great one). Earlier this month I visited and presented 4+ hours of content covering VS 2010, ASP.NET 4 and ASP.NET MVC 2. The talks are demo-heavy and show off a ton of new features. NextSlide.com a .NET shop local to Arizona who has a great product for sharing presentations volunteered to record the talks and publish...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • My VS 2010 and ASP.NET 4 Talks Online

    The past 7 years Ive done an annual all day event in Arizona organized by the most excellent Scott Cate (who always does a phenomenal job organizing the event and making it a great one). Earlier this month I visited and presented 4+ hours of content covering VS 2010, ASP.NET 4 and ASP.NET MVC 2. The talks are demo-heavy and show off a ton of new features. NextSlide.com a .NET shop local to Arizona who has a great product for sharing presentations volunteered to record the talks and publish...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Question about server usage, big community platform

    - by Json
    I’m working on a community platform writen in PHP, MySQL. I have some questions about the server usage maybe someone can help me out. The community is based on JQuery with many ajax requests to update content. It makes 5 - 10 AJAX(Json, GET, POST) requests every 5 seconds, the requests fetch user data like user notifications and messages by doing mySQL queries. I wonder how a server will handle this when there are for more than 5000 users online. Then it will be 50.000 requests every 5 seconds, what kind of server you need to handle this? Or maybe even more, when there are 15.000 users online, 150.000 requests every 5 seconds. My webserver have the following specs. Xeon Quad 2048MB 5000GB traffic Will it be good enough, and for how many users? Anyone can help me out or know where to find such information, like make a calculation?

    Read the article

  • Please help for change wallpaper on terminal where image from http (ubuntu 12.04)

    - by Yan Fachmi
    I need to change the background of my desktop in Ubuntu 12. 04 with a command in terminal, in order to make a script with bash. Does anyone know how to do it? but i want the image from internet... i know if i use local image would like this : gsettings set org.gnome.desktop.background picture-uri file:///home/icorner/wallpaper/curr.jpg but if i use something like this wont work gsettings set org.gnome.desktop.background picture-uri http://www.sergiuhelldragoon.com/wp-content/uploads/2012/04/dota_2_wallpaper_1_1280x800_by_zadelim.jpg Please anyone or somebody?... Thanks & regards, Yan Fachmi

    Read the article

  • Google Webmaster Tools Index dropped to Zero [closed]

    - by Brian Anderson
    Earlier this year I rebuilt my website using ZenCart. Immediately I saw a drop in index status from 59 to 0. I then signed up for Google Webmaster Tools and noticed the Index status took a dramatic drop and has never recovered. I have worked to add content and I know I am not done, but have not seen any recovery of this index since. What confuses me is when I look at the sitemap status under Optimization it shows me there are 1239 submitted and 1127 pages indexed. Most of my pages have fallen off page one for relevant search terms and some are as far back as page 7 or 8 where they used to be on the first page. I have made some changes in the past week to robots.txt and sitemap.xml, but have not seen any improvements. Can anyone tell me what might be going on here? My website is andersonpens.net. Thanks! Brian

    Read the article

< Previous Page | 429 430 431 432 433 434 435 436 437 438 439 440  | Next Page >