Search Results

Search found 22992 results on 920 pages for 'custom pages'.

Page 437/920 | < Previous Page | 433 434 435 436 437 438 439 440 441 442 443 444  | Next Page >

  • Problems after upgrading from 10.04 to 12.04

    - by Paul D
    I upgraded to 12.04.1 from 10.04 recently and am running the Gnome Classic (no effects) desktop. I spent a couple of days tweaking the appearance and behaviour but there are two issues that I can't resolve. The mouse cursor frequently disappears, especially when scrolling through web pages or hovering over links/icons with tooltips. Moving the mouse brings the pointer back but it vanishes again almost immediately. Note I don't have unclutter or Parallels installed. The screen no longer fades to black when left idle for too long - it just cuts out. Bad news when watching content online. Any tips greatly appreciated. Cheers...

    Read the article

  • How to direct a Network Solutions domain name to an html website hosted on Google Drive? [on hold]

    - by Air Conditioner
    To begin with, I'd wanted to take advantage of HTML, CSS, and so on to build a website that looks and works just as I'd like it to. I took a look around on how I could make that work, and I soon saw a lifehacker article showing that its possible to host website files on google drive. I then made sure that the folder containing the files was shared publicly throughout the web, and I now have a working 'google drive hosted' domain for the website. However, I did want to have the custom domain, and so I registered one with network solutions. So now, I'm curious on how I should direct my Network Solutions domain to the index.html I'm hosting on google drive. Would anyone have an Idea?

    Read the article

  • JMeter: how to asign a single distinct value from CSV Data Set Config to each thread in thread group?

    - by JohnnyM
    I have to make a load test for a relatively large number of users so I cant realy use User Parameters pre-processor to parametrize each thread with custom user data. I've read that I should use CSV Data Set Config instead. However I run into a problem with how JMeter interprets the input of this Config. Example: I have a thread group of 3 threads and Loop Count:10 with one HTTP request sampler with server www.example.com and path: \${user}. The csv file (bullet is a single line in file) for CSV Data Set Config to extract the user parameter: 1 2 3 4 5 Expected output is that for thread 1-x the path of the request should be: \x. So the output file should consist of 10 samples per thread namely: for thread 1-1 : 10 requests to www.example.com\1 for thread 1-2 : 10 requests to www.example.com\2 for thread 1-3 : 10 requests to www.example.com\3 but instead i get requests to each \1 - \5 and then to EOF. Does anyone know how to achieve the expected effect with CSV Data Set Config in jmeter 2.9?

    Read the article

  • Showing "Failed" for a SharePoint 2010 Timer Job Status

    - by Damon
    I have been working with a bunch of custom timer jobs for last month.  Basically, I'm processing a bunch of SharePoint items from the timer job and since I don't want the job failing because of an error on one item, so I'm handing errors on an item-by-item basis and just continuing on with the next item.  The net result of this, I soon found, is that my timer job actually says it ran successfully even if every single item fails.  So I figured I would just set the "Failed" status on the timer job is anything went wrong so an administrator could see that not all was well. However, I quickly found that there is no way to set a timer job status.  If you want the status to show up as "Failed" then the only way to do it is to throw an exception.  In my case, I just used a flag to store whether or not an error had occurred, and if so the the timer job throws a an exception just before existing to let the status display correctly.

    Read the article

  • Allow Firefox 8.0 access to /proc/[0-9]*/fd/?

    - by cvt76
    Today I discovered that Firefox 8.0 on Ubuntu 11.10 wanted to have read access to /proc/[0-9]*/fd/. I have made a custom usr.bin.firefox apparmor profile so I have spent many hours testing, but Firefox never wanted to have read access to the /proc/[0-9]*/fd/ directory before, and now suddenly it want it every time I browser the web. The orginal apparmor Firefox profile or any af the abstractions do not allow read access to /proc/[0-9]*/fd/ so I do not understand why this is suddenly happening. Does it sound reasonable that Firefox want to read the /proc/[0-9]*/fd/ directory? or is my system defective?

    Read the article

  • Issue with wired network

    - by Shiju Nambiar
    I have ubuntu 12.10 (downgraded from 13.04 when I faced this issue). I am a newbee and don't know much of technical side like commandline on the terminal. I have a wired internet connection and it was working just fine for last few months since I installed ubuntu. my OS got upgraded to 13.04 and it was still fine. Since last 3 days, network shows connected, connection stays for few seconds and then websites stop loading. Connection shows as connected, but pages don't load. I have to reconnect again to access a website and the connection goes off again in say about 10 to 15 seconds. So I use my internet by connecting and disconnecting every few seconds.

    Read the article

  • Event Handling in jQuery

    - by bipinjoshi
    In the previous part of this series I discussed how HTML DOM can be manipulated using jQuery. Now let's proceed with one of the most common task you do in JavaScript - handling events. jQuery streamlines the JavaScript event handling mechanism by providing you a uniform and easy way to wire events with their handlers without bothering about the cross browser issues. As an example we will create a custom context menu (shortcut menu or right click menu) that displays some options specific to our website instead of the normal context menu of the browser.http://www.bipinjoshi.net/articles/fc9bf7fb-4b81-48cf-9b2d-cee10d2de2f9.aspx

    Read the article

  • Template standard controls for an entirely new look and feel

    - by T
    This is the  Ineta Live player without the O’Data Feed.  It is a good example of taking the plain Media Player provided with the Encoder install and re-templating it to make it your own.  It also has a custom scrub control that is added in.  I generally put my tempates in a separate resource file.  On this project, I discovered that I had to include the template at the document level because I needed the ability to attach some code behind to fire change state behaviors.  I could not use the blend xaml behaviors for change state inside the template because the template can’ determine the TargetObject.   Version 1.01 – 6/10/09 (wow how did a week slip by)

    Read the article

  • Does it matter that the TTL is higher than I've been told?

    - by Andy
    Ok, so the title isn't very descriptive, I know. Basically, I'm trying to configure Outlook.com to use my custom domain! I've followed the steps and made the account etc. and now I have the DNS settings to configure from Windows Live. I added the MX Entries and everything last night but Windows Live is still saying I need to prove my ownership of the domain. The only thing I can think of that I had to use a differrent TTL to the one provided because my web hoster will only allow a minimum of four hours, whereas Windows Live told me to configure the TTL as one hour. Would that stop anything? By the way, my web hoster is JustHost (Shared Hosting)

    Read the article

  • wifi works only after connecting through wire

    - by orustam
    I have fresh installed ubuntu 12.04. it is my first ubuntu installation and i'm a bit confused about the network connection. Wifi shows up and connects(at least it shows that the connection is established), but i can't open any pages, i've tried to ping some sites and it fails either. If i try to connect through a wire it works, what is interesting to me is that after i used my wire connection i can use my wifi properly without a wire pluged in. i think it probably has to do with my settings? I tried to find a solution but can figure it out on my own. My Proxy set to none(have applied it system wide) Please help me if you have any clue:)

    Read the article

  • Best practices for launching a new software version

    - by steve
    I rebuilt a web app to replace a version that we have been using for the last 3-4 years. We have a few thousand clients and a few hundred active users per day. The functionality is basically the same. The new version is a little bit faster with a few enhancement features and there are a lot of behind the scenes changes that the clients will never see. The UI is quite different but ultimately much easier to use and navigate. How should I go about having our clients stop using the old system and start using the new one? I am currently putting together a video that will play on the web site as well as within the app. The video will go through the pages and focus on some key changes. I was also thinking about an intro page that will display once the user logs in and explains some of the features.

    Read the article

  • New Endpoint options that enable additional application patterns

    - by kaleidoscope
    The two communication-related capabilities:  a) inter-role communication and b) external endpoints on worker roles enable new application patterns in Windows Azure-hosted services. Inter-role Communication - A common application pattern enabled by this is client-server, where the server could be an application such as a database or a memory cache. External Endpoints on Worker Roles - A common application type enabled by this is a self-hosted Internet-exposed service, such as a custom application server. For further details click on the following link: http://blogs.msdn.com/windowsazure/archive/2009/11/24/new-endpoint-options-enable-additional-application-patterns.aspx   Tinu, O

    Read the article

  • Template standard controls for an entirely new look and feel

    - by T
    This is the  Ineta Live player without the O’Data Feed.  It is a good example of taking the plain Media Player provided with the Encoder install and re-templating it to make it your own.  It also has a custom scrub control that is added in.  I generally put my tempates in a separate resource file.  On this project, I discovered that I had to include the template at the document level because I needed the ability to attach some code behind to fire change state behaviors.  I could not use the blend xaml behaviors for change state inside the template because the template can’ determine the TargetObject. Version 1.01 – 6/10/09 (wow how did a week slip by)

    Read the article

  • Référencement : Google crée un outil pour renier les mauvais liens entrants et rectifier les mauvaises décisions SEO

    Référencement : Google crée un outil pour renier les mauvais liens entrants Et rectifier les mauvaises décisions SEO Google annonce un nouvel outil permettant de désavouer les liens entrants vers son site et se débarrasser des effets néfastes de certaines stratégies de référencement douteuses. La firme souligne que ce service s'adresse aux webmasters notifiés d'une action antispam basée sur des « liens non naturels » pointant vers leur site. Il permettra de chasser ces backlinks et regagner la réputation de ces pages. [IMG]http://4.bp.blogspot.com/-Ni1-6PIqNHk/UH3EKHp0AoI/AAAAAAAAAFY/UB7LI6eWE6M/s1600/img1.png[/IMG] Le « PageRank » ou l'obs...

    Read the article

  • Changing the content of a website completely, and SEO

    - by Sercan
    I have a blog running since like 10 months, which have 300 organic unique visitors daily, and now I will establish an eCommerce website on that domain. That means, I will delete all the content related to the blog. And publish new pages related to eCommerce using a different script. Content of the blog and the topic of eCommerce are also quite different. How should i do this change in terms of SEO? What should I expect in terms of Search Rankings, organic hit?

    Read the article

  • Tracking 502 bad gateway error

    - by dasickle
    I moved my Wordpress site to WP Engine and now I constantly get 502 errors. I spoke with support and they said that its because I have a lot of DB queries. I ran some tests and my frontpage only has 95 queries and page size is about 500kb. Most inner pages are around 60 queries. All queries are very short. Some people tell me its common with WP Engine because they run nginx. Why do I keep getting these errors and is there a way to track how many of them happen on daily basis? P.S. WP Engine log is empty so cant see the 502's there.

    Read the article

  • XNA: Retrieve texture file name during runtime

    - by townsean
    I'm trying to retrieve the names of the texture files (or their locations) on a mesh. I realize that the texture file name information is not preserved when the model is loaded. I've been doing tons of searching and some experimenting but I've been met with no luck. I've gathered that I need to extended the content pipeline and store the file location in somewhere like ModelMeshPart.Tag. My problem is, even when I'm trying to make my own custom processor, I still can't figure out where the texture file name is. :( Any thoughts? Thanks! UPDATE: Okay, so I found something kind of promising. NodeContent.Identity.SourceFilename, only that returns the location of my .X model. When I go down the node tree he is always null. Then there's the ContentItem.Name property. It seems to have names of my mesh, but not my actual texture file names. :(

    Read the article

  • How to CURL and avoid timeout death (Twitter Down) [migrated]

    - by David
    Twitter is down right now, and one of my site's home pages relies on getting data from Twitter (relies is the problem - it should be more of an accessory feature, as it just shows follow count from its feed). Here's the code in question: function socials_Twitter_GetFollowerCount($username) { $method = function () use ($username) { return file_get_contents('https://api.twitter.com/1/users/show.json?screen_name='.$username.'&include_entities=true'); }; $json = cache('bmdtwitter', 3600, $method, false); $json = json_decode($json, true); return intval($json['followers_count']); } What is a good way to make it so if Twitter is down (or not responsive for some reasonable amount of time), our site doesn't appear to be down (I think the timeout maybe defaulting to 30-60 seconds or more).

    Read the article

  • Alternative Web model

    - by Above The Gods
    One of the problems web apps have against native apps, especially on the mobile front, is the constant need to re-download each web page on request. Ultimately, this leads to slower performance. Why if web apps only download new pages if they're actually needed, not because they're simply requested. For example: perhaps the server can store a web page version in a cookie. Every slight change to the page on the server-side changes the version number. Now instead of the browser requesting a new page each time, why not just check the version number and have the server send the page if they're different? If the page similar, the user can just use a cached page. I'm sure browsers doesn't necessarily have to change to accommodate changes to this, correct?

    Read the article

  • Scammers on the lose pretending to be Microsoft

    - by John Paul Cook
    Minutes ago I received a phone call that the caller ID listed as “Out of area”, which I knew was a bad sign. It was difficult to understand the caller because of his very thick accent. He told me that he was from Microsoft and that my computer was throwing a large number of errors and he was calling to help me. He directed me to use Windows R to open a run dialog box, type eventvwr and then look at the Event Viewer. Within Event Viewer, he instructed me to open Custom Views and then open Administrative...(read more)

    Read the article

  • Redirect Google crawler to different robots.txt via .htaccess

    - by user3474818
    I have googled for the answer all day and still couldn't find an answer. I have a virtual subdomain www.static.example.com which is a mirror site of www.example.com. It means I have just one root folder for subdomain and domain aswell. I want to redirect crawlers to different robots.txt file - robots_static.txt when they see .static in url in which I will forbid indexing via /disallow command. I want to do this because I have duplicated content in Google search results. Subdomain is showing the exact same content as the main domain. Does anyone know how could I achieve that crawlers sees robots_static.txt instead of robots.txt? What I have managed to find so far is this: RewriteCond %{HTTP_HOST} ^www.static.*$ [NC] RewriteCond %{THE_REQUEST} ^[A-Z]{3,9}\ /.*robots\.txt.*\ HTTP/ [NC] RewriteRule ^robots\.txt /robots_static.txt [NC,L] but when I check in webmaster tools, it still sees robots.txt as my robots file instead of robots_static.txt, so it crawls and index everything twice. What did I do wrong? Thanks EDIT: This is my .htaccess file ## # @package Joomla # @copyright Copyright (C) 2005 - 2013 Open Source Matters. All rights reserved. # @license GNU General Public License version 2 or later; see LICENSE.txt ## ## # READ THIS COMPLETELY IF YOU CHOOSE TO USE THIS FILE! # # The line just below this section: 'Options +FollowSymLinks' may cause problems # with some server configurations. It is required for use of mod_rewrite, but may already # be set by your server administrator in a way that dissallows changing it in # your .htaccess file. If using it causes your server to error out, comment it out (add # to # beginning of line), reload your site in your browser and test your sef url's. If they work, # it has been set by your server administrator and you do not need it set here. ## ## Can be commented out if causes errors, see notes above. Options +FollowSymLinks ## Mod_rewrite in use. RewriteEngine On RewriteEngine On RewriteCond %{HTTP_HOST} !^www\. RewriteRule ^(.*)$ http://www.%{HTTP_HOST}/$1 [R=301,L] RewriteCond %{HTTP_HOST} ^www.static.*$ [NC] RewriteCond %{THE_REQUEST} ^[A-Z]{3,9}\ /.*robots\.txt.*\ HTTP/ [NC] RewriteRule ^robots\.txt /robots_static.txt [NC,L] ## Begin - Rewrite rules to block out some common exploits. # If you experience problems on your site block out the operations listed below # This attempts to block the most common type of exploit `attempts` to Joomla! # # Block out any script trying to base64_encode data within the URL. RewriteCond %{QUERY_STRING} base64_encode[^(]*\([^)]*\) [OR] # Block out any script that includes a <script> tag in URL. RewriteCond %{QUERY_STRING} (<|%3C)([^s]*s)+cript.*(>|%3E) [NC,OR] # Block out any script trying to set a PHP GLOBALS variable via URL. RewriteCond %{QUERY_STRING} GLOBALS(=|\[|\%[0-9A-Z]{0,2}) [OR] # Block out any script trying to modify a _REQUEST variable via URL. RewriteCond %{QUERY_STRING} _REQUEST(=|\[|\%[0-9A-Z]{0,2}) # Return 403 Forbidden header and show the content of the root homepage RewriteRule .* index.php [F] # ## End - Rewrite rules to block out some common exploits. ## Begin - Custom redirects # # If you need to redirect some pages, or set a canonical non-www to # www redirect (or vice versa), place that code here. Ensure those # redirects use the correct RewriteRule syntax and the [R=301,L] flags. # ## End - Custom redirects ## # Uncomment following line if your webserver's URL # is not directly related to physical file paths. # Update Your Joomla! Directory (just / for root). ## # RewriteBase / RewriteCond %{THE_REQUEST} ^GET.*index\.php [NC] RewriteCond %{THE_REQUEST} !/system/.* RewriteRule (.*?)index\.php/*(.*) /$1$2 [R=301,L] RewriteCond %{THE_REQUEST} ^GET ## Begin - Joomla! core SEF Section. # RewriteRule .* - [E=HTTP_AUTHORIZATION:%{HTTP:Authorization}] # # If the requested path and file is not /index.php and the request # has not already been internally rewritten to the index.php script RewriteCond %{REQUEST_URI} !^/index\.php # and the request is for something within the component folder, # or for the site root, or for an extensionless URL, or the # requested URL ends with one of the listed extensions RewriteCond %{REQUEST_URI} /component/|(/[^.]*|\.(php|html?|feed|pdf|vcf|raw))$ [NC] # and the requested path and file doesn't directly match a physical file RewriteCond %{REQUEST_FILENAME} !-f # and the requested path and file doesn't directly match a physical folder RewriteCond %{REQUEST_FILENAME} !-d # internally rewrite the request to the index.php script RewriteRule .* index.php [L] # ## End - Joomla! core SEF Section. <FilesMatch "\.(ico|pdf|flv|jpg|ttf|jpg|jpeg|png|gif|js|css|swf)$"> Header set Expires "Wed, 15 Apr 2020 20:00:00 GMT" Header set Cache-Control "public" </FilesMatch> <ifModule mod_headers.c> Header set Connection keep-alive </ifModule> ########## Begin - Remove Etags # FileETag none # ########## End - Remove Etags

    Read the article

  • Separating portion of website to its own server

    - by Brett
    So my job is to take the homepage (or maybe I should say "homesite" because it encompasses a few interrelated pages) and drag this onto its own Apache server. The problem I'm having right now is being able to weed out jumbled/bundled files (such as folders of js, css, and other files that i cant even identify) and knowing what is necessary to keep the homesite running. I'm new to this stuff (I'm an intern) so feel free to ask questions if I'm leaving vital information out. What I'm asking of you guys here is basically any pointers or tips you may be able to give me in order to get the job done. I could use some advice from people with a little more experience in web development. btw: This question may appear as though I have not completed any prior research and that is, for the most part, true. But the problem is I really am not sure how to research this. If you guys could throw me some keywords to play with that would really be helpful. Thanks!

    Read the article

  • Finding the try for an except or finally [migrated]

    - by ?s?
    I'm dealing with some code that has fantastically long methods (10k lines!) and some odd use of try-finally and try-except blocks. Some of the latter are long by themselves, and don't always have the try at the start of the method. Obviously I'm trying to refactor the code, but in the meantime just being able to fix a couple of common pathologies would be much easier if I could jump to the start of a block and see what is happening there. When it's 20+ pages away finding it even with the CNPack rainbows is just tedious. I'm using D2010 and have GExperts (with DelForExp), CNPack and DDevExtensions installed, but I can't find anything that lets me jump from the try to the finally or back. Am I missing something? Is there another add-in that I can use that will get me this?

    Read the article

  • June 14th Webcast: Using Personalization in Oracle eAM

    - by Oracle_EBS
    ADVISOR WEBCAST: Using Personalization in Oracle eAMPRODUCT FAMILY: Enterprise Asset Management June 14, 2012 at 11 am ET, 9 am MT, 8 am PT Personalization is the ability within an E-Business Suite instance to make changes to the look and behavior of OA Framework-based pages without programming (and, therefore, upgradeable!).TOPICS WILL INCLUDE: Oracle eAM customers have leveraged Personalization to address many of the following use cases: making a non-mandatory field mandatory on a Work Order adding a new field to the header region entering the specifications for a new field disabling a field to prevent user entry adding Asset Description to Create Work Request Page viewing Asset Hierarchy in Create Work Request Page setting Auto Request Material Flag to yes by default A short, live demonstration (only if applicable) and question and answer period will be included. Oracle Advisor Webcasts are dedicated to building your awareness around our products and services. This session does not replace offerings from Oracle Global Support Services. Current Schedule can be found on Note 740966.1 Post Presentation Recordings can be found on Note 740964.1

    Read the article

  • facebook og:image not working [closed]

    - by zeinab
    When I try to post a comment (share link) on an article from my page which is actually a page I am using as an application in Facebook: https://apps.facebook.com/ids_newsletter/, Facebook chooses the feed thumbnail image of the post feed randomly from my page. I tried using below to put a custom image for the post but nothing affected the post thumbnail. So what should I do? <meta property="og:title" content="IDS June Newsletter" /> <meta property="og:description" content="Check this out" /> <meta property="og:image" content="MYSITEURL/Images/logobig.png" /> <meta property="og:image:type" content="image/png" /> <meta property="og:image:width" content="200" /> <meta property="og:image:height" content="200" />

    Read the article

< Previous Page | 433 434 435 436 437 438 439 440 441 442 443 444  | Next Page >