Search Results

Search found 13889 results on 556 pages for 'results'.

Page 202/556 | < Previous Page | 198 199 200 201 202 203 204 205 206 207 208 209  | Next Page >

  • Is there any problem with using two slashes in the middle of a URL? [closed]

    - by joshuahedlund
    Possible Duplicate: What does the double slash mean in URLs? I'm working on a mod_rewrite URL structure as follows: http://example.com/search/filter1/filter2/filter3/filter4 There are some conditions where it is OK for the first attribute to be blank, but i want to keep the other attributes in the same position. (Otherwise I can't assume that the attribute in the second position represents what I want it to represent.) However this results in some URLs like this: http://example.com/search//filter2/filter3/filter4 This seems to work in all browsers I've tested (Chrome,Firefox,IE9,IE compatible) and I'm not seeing any errors on the server side, so I can't think of any problems in using it. But it just looks wrong and weird to me and I'm not used to seeing it. Are there any potential downsides to using a structure that encourages URLs like this, or any major reasons no one seems to use it? (Everything I search in Google assumes I'm talking about the two slashes after http:)

    Read the article

  • "Never do in code what you can get the SQL server to do well for you" - Is this a recipe for a bad design?

    - by PhonicUK
    It's an idea I've heard repeated in a handful of places. Some more or less acknowledging that once trying to solve a problem purely in SQL exceeds a certain level of complexity you should indeed be handling it in code. The logic behind the idea is that for the large majority of cases, the database engine will do a better job at finding the most efficient way of completing your task than you could in code. Especially when it comes to things like making the results conditional on operations performed on the data. Arguably with modern engines effectively JIT'ing + caching the compiled version of your query it'd make sense on the surface. The question is whether or not leveraging your database engine in this way is inherently bad design practice (and why). The lines become blurred further when all the logic exists inside the database and you're just hitting it via an ORM.

    Read the article

  • ASP.NET Menu, NavBar And Pager Performance Improvements v2010 vol 1

    Check out the improvements weve made to some of our ASP.NET controls in the DXperience v2010.1 release. We changed the rendering of our ASP.NET AJAX Menu, Navigation Pane and Pager controls. The controls now use semantic rendering combined with advanced CSS styles, which results in a dramatic decrease of HTML output, improved performance and a reduction in the servers workload. Also, several of our other ASP.NET controls like the ASPxGridView and ASPxScheduler also benefit because The primary...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • How to minimize the data loss when laying off a programmer?

    - by thursdaysgeek
    I was just laid off and it was the standard process that is used in the US: call the person to talk to personnel, and remove access to the network while that is going on, then have someone help pack, always have someone with the person until they are escorted from the property. That is supposed to keep an unhappy developer from deleting or damaging software or data: to mimimize data loss. However, it still results in a lot of data loss, as all of the work the programmer was working on is dropped: software not checked in is possibly lost, documents not finished are lost, releases in process are slowed down or stopped, and a huge amount of knowledge could be lost. It seems the potential data loss is more than offset by the actual data loss. How can all losses, both potential and actual, be mimimized?

    Read the article

  • What steps should be taken to ensure that an open source database gets ready for production?

    - by I_like_traffic_lights
    I am considering using GridSQL in a production environment. However, I do have some indications that it is not ready. One is that it got excluded by the offering of EnterpriseDB a while ago, and the forums seem to report a few wrong results and relatively severe bugs. The alternatives to GridSQL, however cost around 100.000$ to buy, so I was thinking to utilize some of this money to ensure that GridSQL gets ready for production. At the same time, I could risk spending 50.000$ and months of work on the development of GridSQL, just to discover that the design was flawed and that a complete rewrite is needed. Then I would have to buy the commercial alternatives to GridSQL and the existence of my startup would be at risk. Question What steps would you take to ensure that there is as little risk as possible that the worst case scenario described above would happen? It is unrealistic that I could do much testing nor code review/coding myself (I am also not the best developer), so please describe where to find the guys that would need to do the work.

    Read the article

  • What causes Google Analytics tags to work on some machines but not on others?

    - by Dallas
    The title of this question says it all. I am trying to update my code from the deprecated _getTracker() method to _createTracker(), but am experiencing inconsistent results. I have tried traditional and Asynch methods using a JSP include, but they all have the same result. My pageviews, and others in the office all show up in analytics. I have tried various test cases, but their visits are just not registering at all for them, but are showing up for me. The client has tried on multiple machines, and I have walked through with them step by step, so I know it's not just user error. I know that javascript being turned off will cause the tags not to work, but I am wondering if what else might cause the tags to not be recognized. I would appreciate any and all ideas.

    Read the article

  • Googling query from anywhere

    - by Shagun
    Some times when I am trying out stuff on my terminal, I get some blockade or some error to resolve which I have to copy the error or last message then open my browser and then Google the query. Is it possible for me that I select the message via mouse and when I right click I get a Google it option as I get when I select some text on a browser? PS : I am not asking how to do Google search via terminal or how to browse web via terminal.These were the results I got when I goggled my question. what I want is something more general. Whenever I select a piece of text I should get a "Google me" option with the right click of my mouse(or maybe with pressing of a key or something). I would prefer that the query takes place in a browser and not in a terminal.

    Read the article

  • Remove home page from Google cache

    - by Steve
    The Google Webmaster Tools Remove URL section allows you to specify a page URL to be removed from the index, or cache, or both. However, I want to remove just the home page, which is / I want to remove it from the cache because it is indexed when the "under construction" page was up. This URL is not recognised by the Remove URL section as an individual page. Instead Google assumes you want to remove the entire website from the index. I've specified /index.php and /index.html to be removed from the cache, but this is not the URL listed in the search results for the home page I want removed from the cache.

    Read the article

  • Site loads Extremely slowly over HTTPS; loads perfectly over HTTP - Sudden and new issue

    - by guest234239048
    My business' website suddenly started loading extremely slowly over HTTPS last night and continues through today. The facts: Page loadtime via HTTPS - 2 minutes +, HTTP - 3 seconds max. NO updates were done on any site files in the past 2 weeks. This is on Shared Hosting HTTPS has worked perfectly over the past 9 months then suddenly failed yesterday. Hosting company and SSL issuer say there are no problems on their end. - Searches for others having similar issues return no results, it appears to be just me... Site is primarily run via php/mysql Currently attempted troubleshooting: Tried all major browsers and different versions - same result. Tried 2 separate ISPs - same result. Tried proxies - same result. Tried 3 separate computers - same result. I'm basically at a total loss here. Does anyone know what could cause such a thing to happen? Please help guide troubleshooting.

    Read the article

  • Where in a typical rendering pipeline does visibility and shading occur?

    - by user29163
    I am taking a computer graphics course. The book and the lecture notes are vague on the on the order of flow between the different steps in the rendering process. For example, if we have specified a view in a scene, and then want to perform a projection transformation for that given view, then we have to go through a sequence of transformations. In the end we end up with a normalized "viewcube" ready to be mapped 2D after clipping. But why do we end up with a cube (ie 3D thing), when a projection results in projecting the 3D objects to 2D. (depth information is lost?) The other line of reasoning is that all information further needed is stored within the "cube" and that visibility detection and shading is performed with respect to this cube and then we perform rasterezation.

    Read the article

  • Search Result Organization

    - by Vecta
    I'm creating an AJAX live search on a website I'm working on. Users will select values from a few dropdowns and a list of products will be returned based on what they select. Some possible fields would be: color, model, make, etc. What type of organization of search results do users tend to find most useful? Is it better to lump them all together (alphabatized) or is it more useful to lump them together by make? In the past I've tended to group them by "make" but I'm not concerned that this will continually force some items with a make toward the end of the alphabet always to the bottom of the list. Any tips are greatly appreciated.

    Read the article

  • Release Notes for 8/16/2012

    Below are the release notes for today's deployment. Source Tab Improvements Revamped fork navigation to make it more clear that you are within a fork verses the parent project. Infrastructure Updated CodePlex to ASP.Net MVC 4.0. RTM+1 how's that for a turnaround? Bug Fixes Fixed a set of user experience issues with the in-line diff view experience. Fixed issue with date mismatches between UTC and local time for discussions and issue tracker Fixed issue where new issues would not show up in search results. Changed the default destination branch for pull requests to master. Have ideas on how to improve CodePlex? Please visit our suggestions page! Vote for existing ideas or submit a new one. As always you can reach out to the CodePlex team on Twitter @codeplex or reach me directly @mgroves84

    Read the article

  • Verify SQL Server Stored Procedures are the same on multiple servers

    I have a stored procedure I push down to all of my servers that does a custom backup job and I want to make sure all servers have the same stored procedure. Is there some way to check without going to each server and reviewing the stored procedure? Check out this to find out. New! SQL Monitor 3.0 Red Gate's multi-server performance monitoring and alerting tool gets results from Day One.Simple to install and easy to use – download a free trial today.

    Read the article

  • Google not showing any pages from my site in the index after three months [on hold]

    - by Alex Coisman
    Despite having a sitemap and using Google Webmaster Tools, it has been over 3 months and my site has not been added to the Google index at all. Here's the site: www.famouslefthandedpeople.com As far as I know, I have done everything correctly. However, there must be something I am overlooking that is preventing Google from indexing the site. I do not have a robots.txt file, so allow/disallow isn't the issue. Although the content of the site is sparse, it is original and not duplicated internally or externally so Panda/Penguin should not be a problem. I have reviewed the answers at Why isn't my website in Google search results? and I don't think it applies here. If it matters, I am using WordPress to create the site. What other factors should I be looking at in order to troubleshoot this?

    Read the article

  • App not showing up in Google Play search on app name [on hold]

    - by William Jockusch
    About 30 hours ago, I released an app on Google Play. I am concerned that if you search on the exact app name, it does not show up in the results, even if you click "show more". https://play.google.com/store/search?q=free+graphing+calculator&c=apps It does show up if you put the name in quotes. But that's awfully low discoverability. https://play.google.com/store/search?q=%22free%20graphing%20calculator%22&c=apps Possibly relevant information: I had an earlier version with a different bundle ID. It was up for just an hour or so, and probably never actually visible to users. How can I fix this?

    Read the article

  • How can I prevent my site from being branded a "content farm?"

    - by Fredashay
    I'm building a small social Q&A site. Another Q&A site that I use was recently branded by Google as a content farm and removed from Google results. I know what Wikipedia says is the definition of a content farm (low quality paid articles and spammy text across the page to catch search engines). That other site I use doesn't do those things, so there must be more to it than that. I want to make sure I don't do anything that causes Google to think my Q&A site is a content farm. What should I do, or avoid doing in designing my site layout?

    Read the article

  • How to prevent a search engines from indexing a section of a page?

    - by BrunoLM
    I have many pages with lots of text in it. But I will always have two sections of text and I want to prevent one section from appearing in search results, the other section must be indexed. <p class="please-index-me">text</p> <p class="get-out">never index me please</p> I thought that maybe if I load the "please don't index me text" with Javascript maybe search engines wouldn't look for it. But I am not sure it would work and this is not really nice. I was wondering if there is a way to tell search engines "hey, this text you can't grab, move on". So, is there a way to do it?

    Read the article

  • Weird sound problems after upgrading to 13.10 (macbook)

    - by Atsch
    I have been having a very weird problem with my sound lately: 1)My right channel is extremely disorted with relativly loud noises, tested with several headphones. 2) After a while, wiered noises start coming from my speakers 3) playing lots of sounds at once causes the the output to become really mushed, e.g. in lmms or games. all points to a: a fault in my sound card, or b: a fault in alsa or pulseaudio, but I don't know how to find out. I will now try using the alsa-daily, and edit this away depending on my results after a restart

    Read the article

  • Clean install of 12.04 won't boot from HDD

    - by MarchingKazoo
    Downloaded 12.04.2. Made CD. Machine boots from CD as expected. Installed onto HD. Upon reboot, "Read error" message appears. Upon booting from CD, the HD appears as expected: One partition for Ubuntu, taking up about 12 gig, the rest empty. Lather, rinse, repeat five times. Results of "sudo fdisk -l": Disk /dev/sda: 200.0 GB, 200049647616 bytes 255 heads, 63 sectors/track, 24321 cylinders, total 390721968 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 512 bytes I/O size (minimum/optimal): 512 bytes / 512 bytes Disk identifier: 0x000dfa3e Device Boot Start End Blocks Id System /dev/sda1 * 2048 386596863 193297408 83 Linux /dev/sda2 386598910 390721535 2061313 5 Extended /dev/sda5 386598912 390721535 2061312 82 Linux swap / Solaris Should this be complicate? Seems no.

    Read the article

  • Disposing of ContentManager increases memory usage

    - by Havsmonstret
    I'm trying to wrap my head around how memory management works in XNA 4.0 I've created a screen management test and when I close a screen, the ContentManager created by that screen is unloaded. I have used ANTS Memory Manager to look at how the memory usage is altered when I do this, and it gives me some results which makes me scratch my head. The game starts with loading two textures (435kB and 48,3kB) which puts the usage at about 61MB. Then, when I delete the screen and runs Unload on the ContentManager, the memory usage drops to 56,5MB but instantly after goes up to 64,8MB. Am I doing something wrong or is this usual for XNA games? Do I have to dispose of everything the ContentManager loads seperatly or do I need to do something more to the ContentManager? Thanks in advance!

    Read the article

  • Testing complex compositions

    - by phlipsy
    I have a rather large collection of classes which check and mutate a given data structure. They can be composed via the composition pattern into arbitrarily complex tree-like structures. The final product contains a lot of these composed structures. My question is now: How can I test those? Albeit it is easy to test every single unit of these compositions, it is rather expensive to test the whole compositions in the following sense: Testing the correct layout of the composition-tree results in a huge number of test cases Changes in the compositions result in a very laborious review of every single test case What is the general guideline here?

    Read the article

  • How Microsoft copy the result from user for Bing.

    - by anirudha
    from some days before i read about the problem google show with another search engine Bing who come from Microsoft. well google show that bing copycat and Microsoft show that they not. i am not known matter much more but predict something that maybe true if google are right in what they tell. if MS really copy the search from User then they use Live service who used by all user who use the internet or on the internet. see the snapshot i have in this snapshot they tell that Help improved search results,internet safety and Microsoft software by allowing Microsoft to collect and retain information about your system, the searches you do and the websites you visit. Microsoft will not use this information to personally identify or contact you. Learn more so why they need the research on what user search on internet even the not contact but user their is no reason as well as bing. well their is a example who maybe possible they  use for bing.

    Read the article

  • Using bone joints

    - by raser
    I am trying to save bone joints to a file, and am using this format. I was wondering if anyone could clear up a few questions I have why do I need to provide rotation data for the bone, if I already gave it the location? How do I calculate the rotation of each axis if I have the relative location from the parent joint? ** EDIT ** After doing some more digging, I think that it has something to do with quaternions, so, could someone point me to a good resource on using quaternions for bone joints? ** EDIT AGAIN ** I think I've solved it, but I don't understand how it works. I can't seem to find any google results explaining it. I'd appreciate if anyone could send resources explaining it to me.

    Read the article

  • How can I do individual file encryption on Dropbox?

    - by Scaine
    I'd like to set a single directory inside Dropbox in which files are encrypted on a file-by-file basis. At the moment, I use a 2Mb Truecrypt container inside my Dropbox which I then have to mount manually, access/change the files within, then unmount manually. At that point, the entire 2Mb uploads to Dropbox. This is a pain for a number of reasons : Dropbox sync will only occur when the Truecrypt container is unmounted, because Dropbox only syncs files that aren't locked and mounting a container locks it. A single byte change to one file inside that container results in the whole 2Mb being uploaded again. It doesn't scale - I was originally using a 10Mb container, but obviously the bigger the container, the longer it takes to sync when it's unmounted. I was wondering if I can somehow use LUKS to implement file-by-file encryption to get round the "container" issues.

    Read the article

  • How to improve a single-paged site search result [closed]

    - by Trisism
    Possible Duplicate: How to SEO a Single-Page website I created an online CV of mine a couple of weeks ago and it has had quite a few visits. Now I want to improve the chance it will appear in google search results; however, my web CV is a one-paged site and it contains only internal links (those with hash #) so I can't really submit a sitemap. I could have changed the internal links to normal links to be processed on server-side, but there's no point of doing so. I'm very new to web SEO so I would really appreciate if somebody can show me what should I do with a single-paged site with internal links to be effectively indexed by crawlers.

    Read the article

< Previous Page | 198 199 200 201 202 203 204 205 206 207 208 209  | Next Page >