Search Results

Search found 6392 results on 256 pages for 'reduce duplicate'.

Page 109/256 | < Previous Page | 105 106 107 108 109 110 111 112 113 114 115 116  | Next Page >

  • Google-Calendar offline?

    - by YSN
    How can I have my google-calendars availabe offline on my PC as well? I have been using dropbox to sync my calendars in ics-format between different PC's instead of using google calendar. The disadvantage of google-calendar is, that I can only see my appointments when online. Now I want to sync my calendars with my android device as well, what does not seem to be possible yet. So I am planning to have an online copy of all my appointments on Google-Calendar, in order to sync them with my Android device. At the same time I want to be able to access my appointments offline (have an offline copy of each calendar on my PC's synced via Dropbox). Is there a way to automatically duplicate and sync Coogle-Calendars with Ubuntu to achieve this? Thanks in advance P.S.: This question is a follow-up, also see my related question, please: How to sync calendar with android without google?

    Read the article

  • How do I keep co-worker from writing horrible code? [closed]

    - by Drew H
    Possible Duplicate: How do I approach a coworker about his or her code quality? I can handle the for in.. without the hasOwnProperty filtering. I can handle the blatant disregard for the libraries I've used in the past and just using something else. I can even handle the functions with 25 parameters. But I can't handle this. var trips = new Array(); var flights = new Array(); var passengers = new Array(); var persons = new Array(); var requests = new Array(); I've submitted documents on code style, had code reviews, gave him Douglas Crockford's book, shown him presentations, other peoples githubs, etc. He still show the same horrible Javascript style. How else could I approach this guy? Thanks for any help.

    Read the article

  • Should I have link rel=next & prev on URLs which have query variables?

    - by user21100
    For example, I have link rel prev & next set up on these pages of products: site.com?page=2 site.com?page=3 (this is my preferred structure by the way and I'm trying to get all the ugly URLs which are littered with query variables deindexed as they are causing duplicate content). So the above URLs are fine but once a filter to narrow product results is selected, like "price", the URL shows like this: site.com?price[1000-1499]=on site.com?page=2&price[1000-1499]=on As of right now, I am having the link rel prev & next dynamically added to the header of these pages but since I am working on getting these query variable URLs pages deindexed, I am wondering if I should get rid of it on these pages? Any thoughts?

    Read the article

  • with dash or not in domain [closed]

    - by menardmam
    Possible Duplicate: Is it better to put hyphens in a domain name ? I have to buy a domain it can be : somebigcompany.com or some-big-company.com I know, it's not a sample answer... but i like to know your point of view... the client what with no dash, i think for Google, and readability the dash is better... i have talk to a expert SEO that sait since panda update of the google algorithm it will be punish to have dash.. i don't believe it ! after your answer, i go to godaddy to buy it thanks in advance

    Read the article

  • How do I render an animation where some frames appear twice?

    - by hustlerinc
    I am animating a sprite. The sprite has 7 different frames, but the animation is 10 frames long. This is because 3 of the original frames appear twice in the animation: 3 -> 4 -> 5 -> 6 -> 4 -> 3 -> 2 -> 1 -> 0 -> 2 Frames 2, 3 and 4 appear twice. This avoids having to store duplicate frames in the spritesheet. How can I render the animation in this sequence with repeated frames?

    Read the article

  • Why do HDFS clusters have only a single NameNode?

    - by grautur
    I'm trying to understand better how Hadoop works, and I'm reading The NameNode is a Single Point of Failure for the HDFS Cluster. HDFS is not currently a High Availability system. When the NameNode goes down, the file system goes offline. There is an optional SecondaryNameNode that can be hosted on a separate machine. It only creates checkpoints of the namespace by merging the edits file into the fsimage file and does not provide any real redundancy. Hadoop 0.21+ has a BackupNameNode that is part of a plan to have an HA name service, but it needs active contributions from the people who want it (i.e. you) to make it Highly Available. from http://wiki.apache.org/hadoop/NameNode So why is the NameNode a single point of failure? What is bad or difficult about having a complete duplicate of the NameNode running as well?

    Read the article

  • Techniques for Working Without a Debugger [closed]

    - by ashes999
    Possible Duplicate: How to effectively do manual debugging? Programming in a debugger is ideal. When I say a debugger, I mean something that will allow you to: Pause execution in the middle of some code (like a VM) Inspect variable values Optionally set variable values and call methods Unfortunately, we're not always blessed to work in environments that have debuggers. This can be for reasons such as: Debugger is too too too slow (Flash circa Flash 8) Interpreted language (Ruby, PHP) Scripting language (eg. inside RPG Maker XP) My question is, what is an effective way to debug without a debugger? The old method of "interleave code with print statements" is time-consuming and not sufficient.

    Read the article

  • List of freely available SEO tools (software) for keyword rank checking? [closed]

    - by Craig
    Possible Duplicate: can anyone reccommend a Google SERP tracker? Requirements: Analysis of site positions on the list of keywords in different search engines; Track keyword positions on search engines. I want see if my keyword rankings have moved up or down; Creating reports. I use Excel + Rank Checker addon for Firefox, to analyze the position of the site in search engines for my keyword list. Are there any tools which tested and working properly. Thanks.

    Read the article

  • How do I deal with content scrapers? [closed]

    - by aem
    Possible Duplicate: How to protect SHTML pages from crawlers/spiders/scrapers? My Heroku (Bamboo) app has been getting a bunch of hits from a scraper identifying itself as GSLFBot. Googling for that name produces various results of people who've concluded that it doesn't respect robots.txt (eg, http://www.0sw.com/archives/96). I'm considering updating my app to have a list of banned user-agents, and serving all requests from those user-agents a 400 or similar and adding GSLFBot to that list. Is that an effective technique, and if not what should I do instead? (As a side note, it seems weird to have an abusive scraper with a distinctive user-agent.)

    Read the article

  • If-else statement comments [closed]

    - by Jin35
    Possible Duplicate: What is a good way to comment if-else-clauses? What is the best way to write comments for if-else statement? There is possible ways: A. //first comment if (condition) { ... } //second comment else { ... } B. if (condition) { //first comment ... } else { //second comment ... } C. if (condition) { //first comment ... } else { //second comment ... } Which one is the best, or there is some better possibilities?

    Read the article

  • Using 301 Redirects on new site when access to old site denied?

    - by Cape Cod Gunny
    I have a situation where I'm standing up a new website on a different web host. I've been denied access to the old site by the hosting company and the old site will most likely be turned off very soon. If my new site contains pages that are named slightly different how do I go about setting up 301 redirects on my new site? For example: www.oldsite.com\aboutus\ www.newsite.com\aboutus.html www.newsite.com\productx.html www.oldsite.com\productx\ Edit: Clarification: The old domain name is different from the new domain name. On my newsite do I just duplicate every page that existed on the old site and place redirect code inside those pages? What does the redirect code look like?

    Read the article

  • How can I backup files on /etc using Deja Dup? [closed]

    - by Chris Good
    Possible Duplicate: How do I open Deja Dup as root? I am using Ubuntu 12.04 LTS. I'm trying to setup Deja Dup to backup the files for several users (/home) and /etc but as it runs as myself, there are many files it cannot read and therefore cannot backup. I want to try running Deja Dup as root as then it should be able to read and backup all files. I've seen some other info about how to set up a particular App to run as root, so I guess I should try that.

    Read the article

  • How are typical users expected to read the documentation in /usr/share/doc?

    - by ændrük
    I only recently learned that there is a huge pile of documentation in /usr/share/doc. How on earth is a typical user supposed to find out about that? It seems like much of it is gzipped, and inaccessible with administrative privileges: $ gunzip examples/letter.tex.gz gzip: examples/letter.tex: Permission denied Are users expected to duplicate each item in their home directory just to read it, or is there a less tedious solution? This arrangement hardly seems conducive to regular browsing. How do normal people read this documentation?

    Read the article

  • Verify uniqueness of new content

    - by rogerkk
    I'm working on a review site, where there is a minor issue with almost duplicate reviews across items. Just a few words are changed. It would be very nice to be able to uncover these duplicates before they are approved by a moderator, and I'm hoping someone could chime in on the best strategy to get there. The site is running Ruby on Rails on a Postgres database and using Thinking Sphinx for search (all on Heroku), and so far the best option I see is to be pulling all the reviews out of the db and using a module like amatch to compare the strings. Not very efficient, so in this case I guess I'll have to limit the number/age of reviews to scan for dupes. Anyone got a better idea?

    Read the article

  • Alternative to "inheritance versus composition?" [closed]

    - by Frank
    Possible Duplicate: Where does this concept of “favor composition over inheritance” come from? I have colleagues at work who claim that "Inheritance is an anti-pattern" and want to use composition systematically instead, except in (rare, according to them) cases where inheritance is really the best way to go. I want to suggest an alternative where we continue using inheritance, but it is strictly forbidden (enforced by code reviews) to use anything but public members of base classes in derived classes. For a case where we don't need to swap components of a class at runtime (static inheritance), would that be equivalent enough to composition? Or am I forgetting some other important aspect of composition?

    Read the article

  • Multiple Java EE Agents on Single Managed Server

    - by tina.wang
    A default JEE agent is created when you create domain, which is named as OracleDIAgent. 1. In Studio, duplicate the agent, change its name to genAgent, change the web application context to genagent. 2: Go to datasource of genAgent, drop all datasources.3: Generate server template. put the jar file under odi\common\templates\wls 4: Deploy this template by update the existing domain. Bring up the config.cmd, choose update existing domain. 5: Update the domain using the template that just generated. Go through the Configuration wizard. (I did not modify anything or configure anything here). 6: The wizard will give information says the deployment was successful. 7: Bring up the admin server and ODI_server1. 

    Read the article

  • Non-unique display names?

    - by Davy8
    I know of at least big title game (Starcraft II) that doesn't require unique display names, so it would seem like it can work in at least some circumstance. Under what situations does allowing non-unique display names work well? When does it not work well? Does it come down to whether or not impersonation of someone else is a problem? The reasons I believe it works for Starcraft II is that there isn't any kind of in-game trading of virtual goods and other than "for kicks" there isn't much incentive to impersonate someone else in the game. There's also ladder rankings so even trying to impersonate a pro is easily detectable unless you're on a similar skill level. What are some other cases where it makes sense to specifically allow or disallow duplicate display names? (I have no idea what to tag this as. I went with game-design because I needed at least 1 tag and I don't have rep to create new ones yet.)

    Read the article

  • How do I remove these errors from my blog so as to get adsense approved?

    - by Serenity
    This is the question I asked on SO site earlier, but didn't get satisfactory replies. hoping to find a solution here.. http://stackoverflow.com/questions/12136796/how-can-i-detect-and-correct-these-errors-on-my-blog/12136829#comment16235061_12136829 In web master tools, apart from the errors in the question link above, it is showing a site map error too as in the screenshot below:- Need guidance please...thanks :) Edit -1 EDIT 2 I had 2 SEO plugins on my blog and I would put meta description for each of my article in both plugins that are All in One SEO and Yoast's "Wordpress SEO". Now I removed all article's meta descriptions from "All in one SEO" the other day but STILL web master tool is showing duplicate meta tags and descriptions. Why??

    Read the article

  • How to protect own software from copying [closed]

    - by Zzz
    Possible Duplicate: How do you prevent the piracy of your software? Is possible to protect some file from copying if you are administrator of machine? I heard some story about some behavior: one software developer sells his software in some way. He is installing it on every client's computer and this software does not work on other computers or cannot be copied physically. How to implement the first and second protection. Is it effectively protection if software costs about $100 for all copies across client's company?

    Read the article

  • Not able to create screenshot of when menu is active [closed]

    - by gotqn
    Possible Duplicate: Screenshot of active window doesn’t work I am using Ubuntu 12.04 and I am searching for a program that will let me to do screenshot of my menus too. Also, it will be very useful if I am able to "cut" only part of the new screenshot and save it as a new image. At this moment, when a menu is shown (for example I click on my clock and the calendar is shown) the "print screen" button is not working. This happens to with applications launchers menus in the left navigation panel. On windows I am using Gadwin Pro. Is there any alternative on Linux?

    Read the article

  • Best book for learning C++? [closed]

    - by gablin
    Possible Duplicate: Is there a good book to grok C++? I'm lacking a good book for learning C++. Although I'm not an entirely novice, I've learnt C++ mostly by trail-and-error and googling so I don't know all about best practices concerning C++. So what book or books should I get and read? Please only put 1 book per answer, for voting purposes, and maybe a short description of who it's targeted at. Hey, what happened to the community wiki option...?

    Read the article

  • "A hard disk may be failing" , but no additional info via gdu-notification-deamon (suggestions?)

    - by blunders
    Got this message after logging in: A hard disk may be failing One or more hard disk report health problems. Click the icon to get more information. On WUBI and 10.04, there was no icon to click, and clicking on made the message go away. After rebooting, the message did not display again. I've got everything on all my hard drives in duplicate, so not super worried about a disk failure, though I am wondering why the message had no info on which disk it thought had problems. Suggestions?

    Read the article

  • Looking for a non-cryptographic hash function that returns a single character

    - by makerofthings7
    Suppose I have a dictionary of ASCII words stored in uppercase. I also want to save those words into separate files so that the total word count of each file is approximately the same. By simply looking at the word I need to know which file it should be in (if it's there at all). Duplicate words should go into the same file and overwrite the last one. My first attempt at solving this problem is to use .NET's object.GetHashCode() function and .Trim() to get one of the "random" characters that pop up. I asked a similar question here If I only use one character of object.GetHashCode() I would get a hash code character of A..Z or 0..9. However saving the result of GetHashCode to disk is a no-no so I need a substitute. Question: What algorithm (or subset of an algorithm) is appropriate for pigeonholing strings into a single character or range of characters (Like hex 0..F offers 16 chars)? Real world usage: I'll use this answer to modify the Partition key used in Azure Table storage as described here

    Read the article

  • Will adding q&a help my site's rankings, and if so, what are the implications of a sub-domain for q&a rather than a path on the site? [closed]

    - by ElHaix
    Possible Duplicate: Subdomain versus subdirectory One of our web properties is doing quite well without any additional links being created on the site, and our link inventory is tightly managed - no user-generated links. To introduce a community aspect to the site, we want to implement a q&a forum. Once in place, new links will populate our link inventory with keywords that are not necessarily targeted to the site. With the q&a on a sub-domain, would that not affect the main site's rankings? What's the best approach for this?

    Read the article

  • Google indexed site's address by accident. What do I do now?

    - by AndrejaKo
    I was making a site for a friend of mine and he wanted to be able to see my progress as I worked on the site, so I decided to put the site on a server on my computer and enable access by a domain name registered to me. It turns out that I forgot to set up a robots.txt file for the site and somehow Google indexed the site. My question is: What do I do now? As I understand it, Google doesn't like duplicate content and my friend could have problems when I upload the new site to his server. Right now his current site, which only has a work in progress page, is first on Google when searching for relevant keywords and I really really don't want to damage that. Is there anything else I need to be concerned about?

    Read the article

< Previous Page | 105 106 107 108 109 110 111 112 113 114 115 116  | Next Page >