Search Results

Search found 29495 results on 1180 pages for 'cross site scripting'.

Page 640/1180 | < Previous Page | 636 637 638 639 640 641 642 643 644 645 646 647  | Next Page >

  • How do I install pgAdmin III for postgreSQL 9.2?

    - by Vector
    I have a Windows server that runs postgresql 9.2. I want to hit it using pgAdmin III from my Ubuntu 12.10 workstation box. I installed pgAdmin III from synaptic and also tried direct download from postgreSQL site using software installer. Regardless, I can get only get pgAdmin III for postgresql 9.1. When I run pgAdmin III and point to my server I get an error message telling me that the database is 9.2 and my pgAdmin III is for 9.1, isn't compatible with 9.2. I can access the server itself fine OK from the Ubuntu box - I have Python programs that hit the database with no problems - but I need pgAdmin III for 9.2 running under Ubuntu 12.10. Is it available? Where do I get it?

    Read the article

  • Did You Know? I'm doing 3 more online seminars with SSWUG!

    - by Kalen Delaney
    As I told you in April , I recorded two more seminars with Stephen Wynkoop, on aspects of Query Processing. The first one will be broadcast on June 30 and the second on August 27. In between, we'll broadcast my Index Internals seminar, on July 23. Workshops can be replayed for up to a week after the broadcast, and you can even buy a DVD of the workshop. You can get more details by clicking on the workshop name, below, or check out the announcement on the SSWUG site at http://www.sswug.org/editorials/default.aspx?id=1948...(read more)

    Read the article

  • What options are there for integrating with payment gateways? [closed]

    - by Rowland Shaw
    It seems that there are only two types of payment gateway service out there at the moment; Either that the entire cart logic is handled offsite (with something like Paypal's Standard option) or the other option being that you need to go through the certification for handling credit card numbers and doing pretty much everything yourself. Ideally, for the project I'm working on, I'm after a bit of middle ground such that I can handle the cart on-site, and only pass over to a payment gateway (with an order amount, billing & delivery details, and order ref) for them to handle the card details, before passing back. I'm sure that I've used e-commerce sites using this pattern before, but I cannot find any payment providers out there that offer this sort of option, so are there any? The only over requirement we have at present is that it must accept orders in Sterling.

    Read the article

  • Accented characters representation in the URL

    - by Dan
    We have support for various languages in our website, including Spanish, French and Swedish. For now, the links in the site are NOT encoded before sent to the browser, sending the real accented chars (if such exists, i.e. href="www.(dot)example(dot)com/héllo.html") and not their HEX representation. This works & looks good on all browsers, including Chrome, FF and IE. However, we care great deal about SEO. We got this tip that encoding the links before sending them to the browser (so instead of linking to http://www.example.com/héllo, we will link to http://www.example.com/h%E9llo) will improve the way search engines will 'understand' the links and the keywords in the URL. This involves some work at our side, so we wanted to know if there's truth in that tip, but couldn't find anything addressing this issue.

    Read the article

  • What is Pseudocode?

    - by Jae
    I've seen a lot of mentions of Pseudocode lately, on this site and others. But I don't get it: What is Pseudocode? For example, the Wikipedia article below says "It uses the structural conventions of a programming language, but is intended for human reading rather than machine reading." Does this mean that it isn't actually used to make programs? Why is it used? How is it used? Is it considered a Programming Language? See the above Wikipedia quote. Is it commonly known/used? Anything else... I honestly don't know where to start with this. I have Googled it and I've seen the Wikipedia article on the topic, but I still don't fully understand what it is.

    Read the article

  • Windows 7 Display Color Calibration Wizard

    Windows 7 comes complete with plenty of new features for you to get the most out of your PC. While some of them may be obvious the operating system has some tools that can help make your life easier. This series will discuss those tools and how to use them. To kick things off let s take a look at the Display Color Calibration Wizard.... Comcast? Business Class - Official Site Learn About Comcast Small Business Services. Best in Phone, TV & Internet.

    Read the article

  • Fonts totally unreadable

    - by user11181
    I've just upgraded from 10.04 to 10.10 and my fonts are all just small rectangles with nothing in them. They are completely unreadable. EVEN THE FONTS IN THE TERMINAL. Thank god the fonts in my Chromium were ok, because I wouldn't be able to find this site or write this message. So far I've tried to use the Appearance - Font thing and it doesn't work. For some reason there was an error with the fonts during the installation. Please help, this is an utter nightmare...

    Read the article

  • F# &ndash; It&rsquo;s time to grow up baby.

    - by MarkPearl
    In the last few months since I started learning F# I have begin to notice an increase in the number of people blogging about the language. Sure, it could just be that I am noticing it more because I am actively looking out for questions and blog posts, but even in my day to day reading of Code Project Daily News and Stack Overflow questions there seems to be increased activity around the language. So what sparked this post? Well, today today I logged in and saw that the latest podcast by DNR was on F# and then immediately afterwards I received an email from CodeProject with a great article comparing F# and Scala. Currently, as of this blog posting (21 May 2010) F# is ranked on the tiobe site at position 43, but I am willing to put money on it that in the next few tiobe ratings this ranking will continue to rise till F# will be a top 20.

    Read the article

  • Ubuntu 13.04/13.10 - low graphics mode/black screen after login

    - by Richard Stokes
    Since last week, my Ubuntu 13.04 installation started booting up in low graphics mode out of nowhere. I tried a number of solutions found on this site but to no avail. The other night, I decided to upgrade to 13.10 to see if the issue would be fixed in the process. Since then, the system starts fine, but when I login to my account, I'm just presented with a black screen. I switched to the terminal and did any necessary updates, but to no avail. I don't even know where to start debugging this issue, so I'd appreciate if someone could point me towards the first place to start looking?

    Read the article

  • How do I stop Google indexing my main page as https [duplicate]

    - by user2897488
    This question already has an answer here: https:// search results appearing on Google for purely http:// site 2 answers Due to historic reasons, we have things set up so that "www.mydomain.com" redirects to "store.mydomain.com". This has worked perfectly fine until recently, when Google appears to be sending visitors to "https:// www.mydomain.com" which doesn't have an SSL-certificate (and never has). Strangely, its only the first link that goes to "https:// www.mydomain.com", all other links point correctly to "http:// store.mydomain.com". Because there is no certificate on the "www" version, users are getting an error message. How do I make Google revert to pointing the main link at "http:// store.mydomain.com" (or even "http:// www.mydomain.com.") If I remove "https:// www.mydomain.com" from Google webmaster tools, will this also remove the redirected page ("http:// store.mydomain.com)? Thanks.

    Read the article

  • How do I get VirtualBox to work?

    - by Karasu
    I can't seem to make VirtualBox to work, I've installed it and reinstalled by 3 ways: Terminal, Ubuntu Software Center and VirtualBox's official web site installation and I keep getting the same kernel error. I found a "solution" to this problem by typing in terminal: sudo apt-get install virtualbox-dkms But it tells my that "virtualbox-dkms is already the newest version". Then I typed in terminal what the error tells me: sudo /etc/init.d/vboxdrv setup but then I tells me that DKMS is actually NOT installed (which is a contradiction). If anyone has encountered the same problem and actually solved it please tell me and explain me step by step 'cause I'm new to Linux.

    Read the article

  • Confused with X.ORG version number

    - by caligula
    There is my trouble: I'm trying to install ATI drivers on my Dell Vostro 3350 laptop. On the AMD's site I followed steps to determine which driver should I use. After selecting what hardware I use I got a link to download amd-driver-installer-12-3-x86.x86_64.run and below was a description of that driver that it was compatible with X.Org version from 6.7 to 7.6. And I do not know what version of X.Org my Ubuntu uses. Cause after googling I found that version is 7.6 in Ubuntu 11.10. But after typing: X -version I got the output: X.Org X Server 1.10.4 Release Date: 2011-08-19 X Protocol Version 11, Revision 0 So what is the exact version of X.Org in Ubuntu 11.10 and how should one reliably determine it? Can anyone construe? Thanks.

    Read the article

  • Images not indexed by google since moving to cdn

    - by dfunkydog
    Last week I moved all the images on coffeeandvanilla.com to a cdn( maxcdn.coffeeandvanilla.com ). The problem I'm having is that although the sitemap—generated by yoast wordpress seo plugin—points images to the correct location, google only indexes[sic] images from the category and page site maps but 0 images from the posts sitemap( see screenshot https://dl.dropbox.com/u/4635252/sitemap.png ) This website has been doing quite well with google image-search before the change, visits from google image search have dropped from ~200/day to 11 yesterday Here is an example entry from the generated posts.xml sitemap http://pastebin.com/vcMRf9VW Can anyone suggest where the problem lies? Why have I lost all my google image juice? Should I just wait some more, how long before really worrying?

    Read the article

  • Is there a way I can filter traffic by page-type based upon URL structure in Google-Analytics or Google Webmaster Tools?

    - by Felix
    I have a local business directory site. I'm trying to segment my incoming traffic by page-type such that I can find out what percentage of traffic is going to zip code pages exclusively and what percentage is going to city/state level pages. I basically want to filter by URL structure to find out what percentage of total traffic zip code pages account for. The reason for doing this is to find out if Google Tag Manager can help with this? Here are the two URL paths: http://www.example.com/ny/new-york/10011/ http://www.example.com/ny/new-york

    Read the article

  • Tab Sweep - Devoxx questions, GlassFish Rest, APAC Java, Lift, JEPs, tools, ...

    - by alexismp
    Recent Tips and News on Java, Java EE 6, GlassFish & more : • Submit a question for Devoxx 2011 Keynotes (Moderator questions) • Devoxx for Java developers (The Java blog) • GlassFish REST Client: ComplexExample.java (Jason) • Oracle Technology Network site for Asia-Pacific developers (OTN APAC) • Notes on deploying Lift apps to GlassFish (Antonio) • Using JSR-250's @PostConstruct Annotation to Replace Spring's InitializingBean (DZone) • The future is in the JEPs (Stephen) • Comparison of Eclipse 3.6 and IntelliJ IDEA 10.5: Pros and Cons (Dzone)

    Read the article

  • DbNetSuite 4.2 released with integrated database charting

    DbNetLink is pleased to announce the release of DbNetSuite 4.2. DbNetSuite is a set of lightweight integrated AJAX enabled ASP.NET components for creating web applications that interface with the database and file systems. New in 4.2 "Out of the box" integration with MS Chart Control Save runtime user configuration with User Profiles Improved performance   Visit the web site for further information, online demos and a fully functional evaluation download. http://www.dbnetsuite.com...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Do Not Track feature of IE10

    - by Pete Herbert Penito
    One of our clients is getting a bit worried about the new "Do Not Track" feature of Internet Explorer 10. Her site is heavily dependent on php sessions (as I imagine many other sites are). This was what she was reading: http://www.bbc.co.uk/news/technology-18288710 I need some clarification, will this affect how sessions (or cookies) work on normal web sites that use the PHP $_SESSION array? Or is it regarding only how advertising works (engadget's article seems to insinuate this)? Can anyone provide a more technical overview (and the ramifications) of PHP-powered websites?

    Read the article

  • Free Training - Building Silverlight Business Applications

    We recently released a new free Silverlight 4 training kit that walks you through building business applications with Silverlight 4. You can also download the entire offline version of the kite here.  You can use the 8 modules, 25 videos, and several hands on labs online or offline from links on the Channel 9 site. Ive included a breakdown and links to all of the content here in this post. The key to this training material is not the features it covers (though it covers a variety of topics including...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Gzip compress offline?

    - by shoosh
    I've configured my site to serve compressed content by putting this line in .htaccess AddOutputFilterByType DEFLATE text/html text/plain text/xml text/javascript text/css application/javascript application/json This works perfectly for almost all files except a few large JSON files that are above 200Kb. For some reason they are not being compressed. I see that they don't using the net tab in firebug and the Network section in chrome. So as a workaround I thought I could compress these files offline and have Apache read them compressed. What tool should I use to compress them? is the linux gzip the one? any special flags or something I should use? What should I put in .htaccess so that the server would know to serve these files with content-encoding gzip ?

    Read the article

  • Google Sites page never shows up in Google Search organic results?

    - by gus
    I use Google Sites (i.e.: https://sites.google.com/site/EXAMPLE/ ) as a convenient way to maintain up-to-date info on several residential properties, info that's often requested by my property agents, its been around for about 1 year, but I still can never get it to appear in organic Google search results or Bing, even if I search the specific keywords such as the street names. I submitted the URL manually to search engines, knowing that my Sites page probably has very few incoming links. Is this expected behavior? The content of my page has simple formatted text, and outgoing links to Picasa/G+/imgur photo albums. Am I doing something wrong or do all GoogleSites pages have poor organic search rank? Thank you very much.

    Read the article

  • The performance implications of IEnumerable vs. IQueryable

    It all started innocently enough. I was implementing a "Older Posts/Newer Posts" feature for my new web site and was writing code like this:IEnumerable<Post> FilterByCategory(IEnumerable<Post> posts, string category) {  if( !string.IsNullOrEmpty(category) ) { return posts.Where(p => p.Category.Contains(category)); }}...  var posts = FilterByCategory(db.Posts, category);  int count = posts.Count();... The "db" was an EF object context object, but it could just as...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • The performance implications of IEnumerable vs. IQueryable

    It all started innocently enough. I was implementing a "Older Posts/Newer Posts" feature for my new web site and was writing code like this:IEnumerable<Post> FilterByCategory(IEnumerable<Post> posts, string category) {  if( !string.IsNullOrEmpty(category) ) { return posts.Where(p => p.Category.Contains(category)); }}...  var posts = FilterByCategory(db.Posts, category);  int count = posts.Count();... The "db" was an EF object context object, but it could just as...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • PHP OOP: Am i following right way?

    - by sineverba
    I'm learning OOP (PHP). I've realized my own CRUD Class, that performs some kind of queries SQL. Btw, a Gasoline asked us to realize a smart, simple web-app where he can update prices of his gasoline (gasoline, diesel, lpg) and via an API i could recall them and display in his site. So, I did create a new Class Gasoline but it perform some methods of CRUD Class public function getPrezzoBenzina($id) { $prezzo_benzina = $this->distributore->sql('SELECT prezzo_benzina FROM prezzi WHERE id = '.$id); return $prezzo_benzina } And so on (code is pseudocode, just to explain). I could perform all my code only with help of Crud Class... without necessity of Class Gasoline. So, what I'm missing about OOP? Where am I wrong?

    Read the article

  • Is it really a security problem to have non secure assets on an ssl page?

    - by blockhead
    My understanding is that this is just an example of being overly cautious, but if my checkout form contains an unsecure asset on it, that doesn't endanger anybody's credit card numbers from being caught by a man-in-the-middle. I'm asking this because every once in while, maybe because of cached content or whatnot, somebody writes in saying that they are seeing this "error" (even though there are no unsecure assets on my page), but they want an explanation. So yes, I can tell all about encryption and certificates and trust and men-in-the-middle. But what do I tell them about this. How do I convince them that the site is 100% safe (and if it isn't let me know that I'm mistaken!)

    Read the article

  • Is there any advantage/disadvantage to using robots.txt to disallow access to legal pages such as terms, privacy policy, etc.?

    - by CaptainCodeman
    As I understand, having repetitive content is a detriment to search engine placement. Given that many websites that use similar or even identical "Terms and Conditions" and "Privacy Policy" pages due to similar legal wording or due to copy & pasting from the same source, would it be a good idea to disallow access to these pages via robots.txt, in order to avoid being penalized for "non-original content"? Or, on the contrary, could the search engines identify this as circumvention and penalize the site for trying to hide content? Or does it not matter?

    Read the article

< Previous Page | 636 637 638 639 640 641 642 643 644 645 646 647  | Next Page >