Search Results

Search found 25614 results on 1025 pages for 'content filter'.

Page 552/1025 | < Previous Page | 548 549 550 551 552 553 554 555 556 557 558 559  | Next Page >

  • Slides for Parallel Debugging windows

    Recently I gave a talk at our Microsoft Shanghai offices on Parallel Programming so I had to update my existing Beta1 deck to Beta2 content. Specifically for Parallel Tasks and Parallel Stacks, I used 5 slides to accompany the demo.In case you are giving talks on parallelism within Visual Studio 2010, please feel free to download and use the updated parallel debugger slides (pptx).TIP: The slides have animations so be sure to F5 the deck for the full benefit and they also have text in the Comments section so be sure to see them at design time too. Comments about this post welcome at the original blog.

    Read the article

  • Job Search Engine Url Structure Issue [closed]

    - by Justin
    Possible Duplicate: What is the best stucture of SEO friendly URL? I am working on a job board, and i'm trying to figure out a good design for URL structure. Some things that I have found through research: 100 - 150 Chars long is ideal 3-5 words in your url, according to Matt Cutts Use .htaccess to force clean urls Do not duplicate data (important) Clean and precise, describing the content Use hyphens On the homepage, I try to detect the users location based on IP, but this isn't always accurate, and not always reliable. So until they put in their city/location, I can't always use this structure but this is potentially work-able. For Searching, a form post to a results page: domain.com/jobs/[city]/[search] ie: domain.com/jobs/toronto/sales manager/ OR domain.com/search/jobs/toronto/sales manager/ or do I remove the word JOBS and just use Search. I trying to keep good search terms in the URL, but also keep it clean and concise. Can someone give me some feedback and thoughts to 'why'...

    Read the article

  • Website Stopped Showing From Google Search Results Sunddenly

    - by Aman Virk
    I have a design and development blog http://www.thetutlage.com (1.5 years old), which was doing really well in Google search as I was getting over 70% of my traffic from Google. Now suddenly from last two days it reduced the amount of traffic from 70% to 20% and also when I am trying to search for the exact posts that I can created even after appending my website name to it does not show any results for that. Sample Search Text: JQuery Game Programming Creating A Ping Pong Game Part 1 I have post with exact same title and it does not show it on Google search anywhere. I am totally shocked, I write my own unique content and follow Google guide lines like bible. Also there is no message under my webmasters account stating any problem or error.

    Read the article

  • Amazon Kindle e-Ink based device programming: Java ME CDC old school

    - by hinkmond
    If you like doing Amazon Kindle development in the old-school way (Java ME CDC-based apps) on their e-Ink based readers, then here's how to download and use the Amazon Kindle Development Kit (KDK). See: Download Amazon KDK Here's a quote: We're excited to introduce the all- new Kindle family: Kindle, Kindle Touch, and [blah-blah]. The KDK has APIs, tools, and documentation to help you create active content for Kindle, Kindle Touch, and other E Ink Kindles. Kickin' old school with Java ME CDC technology is the way to go. You can come up with the next Word with Friends this way. Hinkmond

    Read the article

  • Joomla 2.5 -- Adding a custom field to menu-item-edit-form

    - by philipp
    I would like to add a new Field (Select list of all menu-items) to the menu item-edit form. To do so I was setting up an system plugin with the following directory structure: languageroot/languageroot.php languageroot/form/form.xml As you can see in the posted code, that is all very basic to try out. Only after adding the following lines: <li <?php echo $this-form-getLabel( 'langroot-text', 'main' )? <?php echo $this-form-getInput('langroot-text', 'main' ); ? </li to: /admininstrator/components/com_menus/views/item/tmpl/edit.php a textfield shows up. Is it possible to inject the field without touching the edit.php? Is there anywhere a good tutorial about the JForm api? Is a system-plugin the right kind, or could it be a content plugin, or should it even be a component?

    Read the article

  • Issue in Webscrapping in C# : Downloading and parsing zipped text files

    - by user64094
    I am writing an webscrapper, to do the download content from a website. Traversing to the website/URL, triggers the creation of a temporary URL. This new URL has a zipped text file. This zipped file is to be downloaded and parsed. I have written a scrapper in C# using WebClient and its function - DownloadFileAsync(). The zipped file is read from the designated location on a trapped DownloadFileCompleted event. My issue : The Windows 'Open/Save dialog is triggered". This requires user input and the automation is disrupted. Can you suggest a way to bypass the issue ? I am cool with rewriting the code using any alternate libraries. :) Thanks for reading,

    Read the article

  • Veranstaltungshinweis: 2. Oracle Breakfast

    - by nospam(at)example.com (Joerg Moellenkamp)
    Am 26. April findet das zweite Oracle Oracle Breakfast in Hamburg in der Geschäftsstelle (Kühnehöfe 5) statt: Also Futtern mit technischem Content. Auch diesmal gibt es zwei Vorträge. Agenda 9:30Willkommen zum Frühstück 10:00Solaris 11 im Detail - Einbindung in heterogene Netze (CIFS-Dienst etc.)Joerg Moellenkamp 11:30Kaffeepause 12:00ZFSSA praktischEinbindung einer ZFS SA in heterogene Netze, aber wie? Vortrag & Livedemo unter VirtualBoxDirk Nitschke 13:30geplantes Ende Anmelden könnt Ihr euch mit einer formlosen Mail an [email protected]. Das ist ein Forwarder an die Addresse der Kollegin, die das intern bei uns organisiert, deren Mailaddresse ich nicht unbedingt für Spammer verteilen möchte ...

    Read the article

  • Is it safe to Block These URLs with Robots.txt?

    - by Edgar Quintero
    I have a website that has all URLs optimized and 301 redirected from nasty URLs to clean ones. However, everywhere throughout the site the unclean URLs are linked in menus, content, products, etc. Google currently has all clean URLs indexed, along with a few unclean URLs too. So the site still has linked everywhere the old URLs (ideally this wouldn't be the case but this is how it is ATM). I would like to block the unclean URLs with robots.txt. The question: If I block these unclean URLs with the robots.txt, when the entire website is linked with them (but they all redirect to the clean version), will this affect the indexing status at all?

    Read the article

  • Web App vs Portal Platform - convincing the customer

    - by shinynewbike
    We're evaluating a set of requirements for a customer who wants Liferay which mainly has AAA and Web CMS requirements, and allowing user to upload their own content. Also all inetgration is via web services. However there is no need for other features such as actual "portlets", i18n, mashups, skins, themes, tagging, social presence, no collaboration etc So we feel we can do this as a standard JEE web app and not use Liferay (or any other portal product) since these are overheads we dont need. The customer feels the Web CMS requirements + user upload justify the "portal" product. Can anyone help me with some points to convince the customer? Assuming our point of view is right.

    Read the article

  • How would I pursue a track in front-end web development?

    - by Koviko
    I've recently been put on heavy JavaScript projects and have become fond of the front-end world in comparison to the back-end. I have always been good at proper markup and CSS, and coupled with AJAX, pretty animations, and dynamically generated content, it's become a much more interesting and flashy world for me. I would like to be able to continue to hone my craft in the same way that I was able to become proficient at back-end development with PHP: getting paid to do it. How would I market myself as a front-end web developer with a strong interest in dynamic JavaScript-driven websites? Due to my strong background in back-end development, how would I find the companies that wouldn't waste my front-end skill set on simple HTML/CSS development? And as a bonus, how would I apply this to being a contractor/freelance developer rather than a salaried employee? While I like the idea of being able to remain a part of my creations, I also dislike the maintenance phase of projects.

    Read the article

  • My VS 2010 and ASP.NET 4 Talks Online

    The past 7 years Ive done an annual all day event in Arizona organized by the most excellent Scott Cate (who always does a phenomenal job organizing the event and making it a great one). Earlier this month I visited and presented 4+ hours of content covering VS 2010, ASP.NET 4 and ASP.NET MVC 2. The talks are demo-heavy and show off a ton of new features. NextSlide.com a .NET shop local to Arizona who has a great product for sharing presentations volunteered to record the talks and publish...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • My VS 2010 and ASP.NET 4 Talks Online

    The past 7 years Ive done an annual all day event in Arizona organized by the most excellent Scott Cate (who always does a phenomenal job organizing the event and making it a great one). Earlier this month I visited and presented 4+ hours of content covering VS 2010, ASP.NET 4 and ASP.NET MVC 2. The talks are demo-heavy and show off a ton of new features. NextSlide.com a .NET shop local to Arizona who has a great product for sharing presentations volunteered to record the talks and publish...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • What is the current legal status of magnet links?

    - by Moonwalker
    Prelude: I develop a cloud service which could be described as dropbox meets torrents and as side effect it enables distribution of arbitrary content via magnet links. Certain amount of magnet links will be displayed on the main website (I will be able to remove them one-by-one or ban users but no more). I will not be able to avoid magnets without complete rework of overall project architecture and either way it will hurt overall performance badly, probably making service meaningless. So my question is, what should I do, to avoid legal problems if my site in a nutshell is just a collection of magnet links? Privacy achieved via end-user encryption, so there is almost no access restrictions on the website. And anyway will help me any? Will hosting in particular country help me?

    Read the article

  • How do I grep in a variable?

    - by Ashfame
    How to do a grep in a variable? I have stored the wget output in a variable and I need to extract out some strings from it. Like the content of the variable is upgrade http://wordpress.org/download/ http://wordpress.org/wordpress-3.0.5.zip 3.0.5 en_US 4.3 4.1.2 I need to check if the string contains the word upgrade, so I can do a simple grep and then check the exit status of it by $? and proceed. But how can I get the value 3.0.5 which is actually the fourth word? And how to actually grep in a variable?

    Read the article

  • Please help for change wallpaper on terminal where image from http (ubuntu 12.04)

    - by Yan Fachmi
    I need to change the background of my desktop in Ubuntu 12. 04 with a command in terminal, in order to make a script with bash. Does anyone know how to do it? but i want the image from internet... i know if i use local image would like this : gsettings set org.gnome.desktop.background picture-uri file:///home/icorner/wallpaper/curr.jpg but if i use something like this wont work gsettings set org.gnome.desktop.background picture-uri http://www.sergiuhelldragoon.com/wp-content/uploads/2012/04/dota_2_wallpaper_1_1280x800_by_zadelim.jpg Please anyone or somebody?... Thanks & regards, Yan Fachmi

    Read the article

  • SEO With a Linkwheel

    You will be amazed to know that more than 68 % of internet users do not go farther than first page on search results. That means you have to be among the top page ranks to have any business online. If you are not among those listed on the first page of search results, you are missing out on most of the users searching for products, information and content on Google and other search engines. Linkwheel is a wonderful strategy for search engine optimization (SEO). If done well, your search engine rankings will skyrocket in a very short time.

    Read the article

  • Google Webmaster Tools Index dropped to Zero [closed]

    - by Brian Anderson
    Earlier this year I rebuilt my website using ZenCart. Immediately I saw a drop in index status from 59 to 0. I then signed up for Google Webmaster Tools and noticed the Index status took a dramatic drop and has never recovered. I have worked to add content and I know I am not done, but have not seen any recovery of this index since. What confuses me is when I look at the sitemap status under Optimization it shows me there are 1239 submitted and 1127 pages indexed. Most of my pages have fallen off page one for relevant search terms and some are as far back as page 7 or 8 where they used to be on the first page. I have made some changes in the past week to robots.txt and sitemap.xml, but have not seen any improvements. Can anyone tell me what might be going on here? My website is andersonpens.net. Thanks! Brian

    Read the article

  • How to use rel=canonical with Sitecore aliases?

    - by Mike G
    I have inherited a Sitecore architecture that is a mess from an SEO duplicate content POV. There are multiple aliases that have been created (and indexed by the search engines) for many of the 2nd tier pages of the site. Due to server issues, I am not able to 301 redirect these duped pages, so I would like to use the rel=canonical tag in an attempt to try and get Google/Bing to recognize the correct pages I would like to appear in the index. I have blocked the most extraneous duped pages with a robots.txt file, however, since Google/Bing have already spidered many of the duped pages, I need to keep them accessible to the spiders, BUT removed from the index. The catch is, since the duped pages are aliases (and don't really physically exist in Sitecore that I can find), I am not sure how to go about using rel=canonical - or if I even can in this situation..?

    Read the article

  • Is there any way to discover the traffic of a site I don't control?

    - by George Bailey
    Given the following: The website does not call any external images or scripts, all the content is hosted on a server that is in our control. The website does not contain the meta tag, nor does it contain the html file that would authorize a Google Account access to Webmaster Tools. The access logs have not been provided to any 2nd or 3rd party. Is it possible for a 3rd party to get an idea of how many hits the site is getting, or are they limited to just seeing how high the site ranks? How could the 3rd party determine how well the site is doing under these restrictions? Is there a website for that that you know of?

    Read the article

  • joomla sometimes messes up urls, probably cache involved

    - by Bakaburg
    Is a bit i'm having this problem and i really cannot get the hang of it... Every once in while my joomla site messes up links url and for example from something like this: http://www.sism.org/index.php?option=com_comprofiler&task=userslist&listid=4&Itemid=123 it becomes like this: http://www.sism.org/index.php/component/k2/administrator/components/com_dump/assets/css/images/stories/inrilievo/sism/htm/index.php?option=com_comprofiler&task=userslist&listid=4&Itemid=123 the new page has the right content but there are no css and other linked resources. Usually i solve the problem by deleting all the cache and turning it off and on again. Of course this is pretty annoying especially for my association. Does any one have any clue on this? Watching the URLs the components involved seems to be K2 and Jdump. Thanks

    Read the article

  • My colleague can't visit our website through her provider after long downtime

    - by Peter Westerlund
    We did a frontpage update some days ago that caused the site to crash. The site was down for several hours. After troubleshooting, we concluded that we needed to cache more content. It had been run too many queries. After solving that and rebooting of server, we here in Sweden and Norway were again able to visit the site. But a colleague in Tunisia couldn't. It seems to work from another internet provider but not her own. What could have happened? And what should we do? Edit: I should add: She is able to visit the site through tunnel at anonymouse.org.

    Read the article

  • Grep in a variable

    - by Ashfame
    How to do a grep in a variable? I have stored the wget output in a variable and I need to extract out some strings from it. Like the content of the variable is upgrade http://wordpress.org/download/ http://wordpress.org/wordpress-3.0.5.zip 3.0.5 en_US 4.3 4.1.2 I need to check if the string contains the word upgrade, so I can do a simple grep and then check the exit status of it by $? and proceed. But how can I get the value 3.0.5 which is actually the fourth word? And how to actually grep in a variable?

    Read the article

  • DBA Best Practices: A Blog Series

    - by Argenis
      Introduction After the success of the “Demystifying DBA Best Practices” Pre-Conference that my good friend Robert Davis, a.k.a. SQLSoldier [Blog|Twitter] and I delivered at multiple events, including the PASS Summit 2012, I have decided to blog about some of the topics discussed at the Pre-Con. My thanks go to Robert for agreeing to share this content with the larger SQL Server community. This will be a rather lengthy blog series - and as in the Pre-Con, I expect a lot of interaction and feedback. Make sure you throw in your two cents in the comments section of every blog post. First topic that I’ll be discussing in this blog series: The thing of utmost importance for any Database Administrator: the data. Let’s discuss the importance of backups and a solid restore strategy. Care to share your thoughts on this subject in the comments section below?

    Read the article

  • Ubuntu 12.10 installation Problem in Wubi

    - by HammerEngineer
    I'm trying to install Ubuntu 12.10 on drive "I:\" along my Windows 7 with Wubi. But I got the following error message: Error executing command >>command=C:\Windows\System32\bcdedit.exe/set {0bfbf1c0-11ba-11e2-a8ef-fooae3684131}device partition=I; >>retval=1; >>stdrr=An error has occured setting the element data. the request is not supported. >>stdout= For more information,please see the log file; c:\users\hammer~1\appdata\local\temp\wubi-12.10-rev273.log What should I do now ? @ http://askubuntu.com/users/5592/agmenor is too large for this content area(Exceeds limit). And the the "bcbc" link is same to my problem.But I can't fix it till now. Any suggestion please ?

    Read the article

  • Is Perforce as good at merging as DVCSs?

    - by dukeofgaming
    I've heard that Perforce is very good at merging, I'm guessing this has to do with that it tracks changes in the form of changelists where you can add differences across several files in a single blow. I think this implies Perforce gathers more metadata and therefore has more information to do smarter merging (at least smarter than Subversion, being Perforce centralized). Since this is similar to how Mercurial and Git handle changes (I know DVCSs track content rather than files), I was wondering if somebody knew what were the subtle differences that makes Perforce better or worse than a DVCS like Mercurial or Git.

    Read the article

< Previous Page | 548 549 550 551 552 553 554 555 556 557 558 559  | Next Page >