Search Results

Search found 9724 results on 389 pages for 'pro zeck'.

Page 230/389 | < Previous Page | 226 227 228 229 230 231 232 233 234 235 236 237  | Next Page >

  • Making a calendar tool with Google API. Should I use OAuth Service Account?

    - by Goluxas
    I'm creating a calendar tool for a client. He has his work calendars in a Google account, which I have access to. I'll be using the Google Calendar API for the tool, but I'm new to Google APIs and the OAuth stuff. I'll be using PHP. Do I need to use an OAuth2 Service Account, or should I make a regular OAuth2 Web Application Client ID from that Google account? In case this information is helpful, the tool I will be making will do this: 1) Allow my client to fill out a form which will create a new calendar and populate it with several events. 2) Send out an email to a mailing list twice a week listing changes that have been made to these calendars. I'll also be making a page that displays the iframes for each calendar, so his clients may see them even if they do not have a Google account.

    Read the article

  • Kloxo - Use third party nameserver

    - by LePhleg
    Here's my case: I got my domain from a registrar. I got my VPS up from a provider who also offers free DNS hosting, on their own server (separately from my VPS, IP:111.111.111.111). So I got these third-party nameservers, lets name them: ns1.provider.com ns2.provider.com and pointed my domain registrar to them. I let it propagate for 24 hours so I believe its ok. (if I ping my domain I get the IP for these nameservers - 111.111.111.111) Afterwards I headed to my providers cPanel (the one they gave me for the DNS hosting service) and changed the default A Record IP from 111.111.111.111 to the IP associated with my VPS (IP:222.222.222.222). The last step, as I see it, is to configure my VPS via Kloxo panel and add the domain. Can someone guide me how can I achieve that? Is it just enough to create a new DNS Template and add as Primary DNS "ns1.provider.com" and Secondary DNS "ns2.provider.com"? Every tutorial I found is using Kloxo to create local nameservers on the same VPS which is something I want to avoid to save resources..

    Read the article

  • Framework for interaction between web-page and server-side script

    - by Carrier
    I want to make a web-page that will have several controls elements, among which there are elements like check-boxes, radio-buttons, "range selectors" (one can specify the min and max value, like it is done when you select range for prices in the online markets). The new values shall be sent to the "server-side", once changed (without any Submit buttons etc), and the "server-side" can return something (one or more numbers, etc). Does anyone know a good ajax-like framework that allows (with minimal adaptation / changes) to make such solution in an easy way? It will be good if the server-side of existing solution will be in Perl (not a big deal, but I know it much better than PHP or something else). Set of controls might change and depend on other parameter, so adding one extra element should not cause rewriting the whole thing. P.S.: I haven't working in this area for quite a while, so not aware of existing solutions in this area, and don't want to invent the wheel and write everything from scratch for something that already exist (at least, I hope so). Thanks in advance!

    Read the article

  • will redirect subdomain lose Google SEO links

    - by user29160
    Because of a change in brand, I want to redirect our subdomain.domain.com to a newdomain.com. The content being exactly the same, I was thinking of using a 301 wild card 301 redirect to newdomain.com. I noticed it is not possible to do a redirect in Google webmasters tool as you can with a root domain. Is there a way I can do this redirect without losing all the backlinks referenced with Google? All the best, mark

    Read the article

  • Mobile traffic of my second site redirected to my base site [Edited,have a look]

    - by Faheem Rasheed
    I have a strange issue with my website and it seems i am unable to understand what cauese the problem.I would highly appreciate your help.The Scenario is I have two websites. Website A Website B Website A is simply hosted within the root directory of my hosting account.Within this root *directory* i have a sub folder " subfolder A " and within which is another subfolder " subfolder B " that contains my site Website B so the path looks like root/sublfolder_A/subfolder_B/ that contains my **Website B ( i.e subfolder_B) ** All goes fine.When i access the website B from my desktop/Laptop Website B is loaded normally. but when i access the Website through a mobile device , mobile site of Website A is loaded while as it should load website_B Also , to let you know , both websites have different URL's and not subdomain's or anything. What could be the problem ? htaccess of website B or website A ? or something else ? Here is the htaccess of my website B RewriteEngine On RewriteRule .* index.php [F] RewriteBase / RewriteRule .* - [E=HTTP_AUTHORIZATION:%{HTTP:Authorization}] RewriteCond %{REQUEST_URI} /component/|(/[^.]*|\.(php|html?|feed|pdf|vcf|raw))$ [NC] RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_FILENAME} !-d RewriteRule .* index.php [L]

    Read the article

  • Products missing from backend after import

    - by byronyasgur
    My client imported about 70 products into magneto and the all imported properly it seemed, but now only about half of them appear in the catalog ( backend ) searching for the SKU returns no results. I have searched the net for an hour looking for some information on this but I cant find any, furthermore I cant seem to find out how to verify whether the products are in the database or not, I thought I could just look for the products table in phpmyadmin, but Magento's EAV database structure looks more complicated than I'm used to. Does anyone know how to fix this, or even to locate products in the database. Edit : From talking to the client she thinks that this might have happened after she did a reindex.

    Read the article

  • Alt text vs CSS sprites (SEO vs speed)

    - by leeoniya
    I'm reworking our site to reduce HTTP requests and blocking requests by concatenating JS, css, gzipping, loading all JS via LABjs and using CSS sprites for images that were loaded individually via <img> tags before. Progress has been great so far - 5x page load performance improvement. However, we're in the top 5 organic search ranking in google for many targeted keywords and phrases. I'm afraid eliminating so many img tags with alt attributes can hurt our SEO. Does anyone have any experience with alt tag manip/removal and effects on SEO positions? Is previous rank "sticky"?

    Read the article

  • Can I use asterisks in URLs?

    - by KajMagnus
    Are there any reasons I shouldn't use an asterisk (*) in a URL? Background: With asterisks, I could provide these nice and user friendly (or what do you think??) URLs: example.com/some/folder/search-phrase* means search for pages with names starting with "search-phrase", located in /some/folder/. example.com/some/**/*search-phrase* means search for any page with "search-phrase" anywhere in its name. example.com/some/folder/* means list all pages in /some/folder/ (rather than showing the /some/folder/index page).

    Read the article

  • Which is the appropriate content-type meta tag value ?

    - by Argoron
    I have a question about the meta tag content-type. When starting to build my site (HTML+PHP+JS), I copied a lot of the meta tags over from elsewhere, and I have, amongst others, the following: <meta http-equiv="content-type" content="application/xhtml+xml; charset=iso-8859-1" /> Now, I've seen that tag is being used a lot with the value "text/html". I've been searching the web but could not find a comprehensive explanation regarding what the difference between both is. The "text/html" intuitively sounds more straightforward to me. Should I change my tag to that, or might the "application/xhtml+xml" be an equivalent solution ? Alternatively, can anyone point me to a resource where the different values for these tags are listed and explained in a clear manner? Thanks in advance

    Read the article

  • How long is the penalty for Duplicate ecommerce content after it has been ressurected

    - by will
    I am fixing all of the duplicate content on my ecommerce site with all orignal descriptions etc. How long does it take google to start ranking it again? I used to have a good ranking that converted quite a few sales, in the last week i have had next to nothing. Also would the disclaimer i created under each product be considered duplicate content because it is on most of my product pages & is the same.

    Read the article

  • Does similar Titles considered duplicates relating to SEO

    - by Uri
    I have built a testing service and I wonder if I should be concerned that search engines will consider similar titles as duplicates. For example: Some URLs title can be differed from others by only one word such as "Senior" and "Junior" Title A: C++ Online Test for Juniors Title B: C++ Online Test for Seniors Another example is with the "+" sign: Title A: C Standard Library Online Test for Seniors Title B: C++ Standard Library Online Test for Seniors Should I assume search engines will understand there is a difference in the titles? And the titles are not duplicated?

    Read the article

  • How to push through a domain transfer in spite of the 60 day rule

    - by corsiKa
    I recently purchased a domain through a registrar which I won't name here. Within the first five minutes of logging in, I found a severe vulnerability that allows me access to all registration details of all users. Simply put, I do not trust this registrar with any kind of business. But I'm unable to transfer the domain because, for some reason, it has to exist in its current state for 60 days. We're planning to launch the site this weekend - we can't wait 60 days. But I can not trust this registrar: if I found such a severe vulnerability in the first few minutes, how many more similar un-trustables will I find in those 60 days? Is there a higher authority to whom I can submit a case to get my domain transferred to a different registrar?

    Read the article

  • Hosted News Carousel

    - by user19240
    I'm trying to add a news carousel on my OneHub site. I know there are quite a few plugins that I can use, but I'm looking for a ready-to-use hosted solution where people can log-on and manage the content, and I can simply include as an iFrame widget on our OneHub site. We would prefer to use Expression Engine, since we're using this on some of our other websites. It looks like Expression Engine + JQuery Cycle Rotator might be an option, not sure how easily I can get this onto OneHub though... WordPress has a carousel plugin as well that I'll explore. Anyhow, while I'm trying out these options, I figured I would put a question out there for any thoughts or advice, other solutions, etc. Any help would be greatly appreciated! Thanks, Alan

    Read the article

  • How do I fix the paths of my website?

    - by EASI
    I have Joomla 2.5.7 in my client's server updated recently from 1.6 to 1.7. I did not make that site but I am responsible for it now. I prefer to make a site from zero. Now users area clicking on the menu options and when joomla send them to a meta-url like http://iap.pa.gov.br/acervo they get the message 404 The requested URL /acervo was not found on this server. Would that be because they moved joomla folder from the root to root/iap (name of the site)? If it is, what is the configuration to adapt to that new folder?

    Read the article

  • How hard is to be the anonymous owner of a website?

    - by silla
    I'd like to create a website with a very radical political message. It won't be unethical (encouraging violence, etc) but I feel the points I plan to list in it will definitely make me a lot of enemies. How hard would it be to protect my identity from anyone finding out who I am? I know domains always have a $10/year option for privatizing your registration information but is there any other protection I should think about having? Thanks!

    Read the article

  • Site inaccessible by some people, fine for others [on hold]

    - by Paul Howell
    A couple of days ago my website www.howellphoto.com (hosted by one.com, wordpress site) started loading really slowly, and I have been unable to access any pages linked from the homepage. Several of my friends have found the same issue, yet many are able to access the site without problem. Live support at one.com have not been all that much help, requesting the ip addresses of a few people who cannot access the site, and saying it could be a firewall issue. Wordpress support (my site was created in prophotoblogs) have been better and have updated all plugins, etc, but can see no issue from their end. My main issue is that even if there was a local fix that I could do on my computer, this would not help wih any potential customers visiting my site for information! This is driving me crazy!!! Any help will be legendary! Cheers, Paul

    Read the article

  • PHP efficiency question [closed]

    - by Ron
    Hello everyone. I am working on website and I am trying to make it fast as much as possible - especially the small things that can make my site a little bit quicker. So, my to my question - I got loop that run 5 times and in each time it echo something, If I'll make variable and the loop will add the text I want to echo into the variable and just in the end I'll echo the variable - will it be faster? loop 1 (with the echo inside the loop) for ($i = 0;$i < 5;$i++) { echo "test"; } loop 2 (with the echo outside [when the loop finish]) $echostr = ""; for ($i = 0;$i < 5;$i++) { $echostr .= "test"; } echo $echostr; I know that loop 2 will increase a bit the file size and therefore the user will have to download more bytes but If I got huge loop will it be better to use second loop or not? Thanks.

    Read the article

  • Somehow Google considers a properly 301'd URL as 200 and is still indexing the new content in old page?

    - by user2178914
    We redirected all the old URL's to new ones properly using htaccess. The problem is Google, somehow is still finding content in the old page(which it shouldn't) and stores it in the cache rather than the new URL. For eg: Old Page- http://www.natures-energies.com/iching.htm New Page- http://www.natures-energies.com/index.php?option=com_content&view=article&id=760 If you type the old URL into the browser it redirects If you fetch the old URL as Googlebot in the webmaster tools the header says 301/permanently redirected. If I try to crawl as any other bot it still says 301 redirected. Even if you click the old link in Google it redirects to the new URL. Only in its cache it shows the old URL and moreover it shows the new content in it! I am stumped on how Google manages to grab the new content and puts in the old URL instead of the new one! One more interesting thing is that if I try a cache for the new page it shows the cache of the new content with old URL! Any help would be appreciated. I am at end of my wits. I think i have tried almost everything. Is there anything that I'm missing to see? You can use this search to find the old url's. Maybe you'll some patterns that i missed. site:www.natures-energies.com inurl:htm -inurl:https|index

    Read the article

  • Include latest searches in search engines index

    - by drcelus
    My websites generally include a page with the (user input) latest searches. I know it's not a good security practice to allow this since you can find undesired content. On the other hand it boosts the number of pages indexed since every new search can provide a link on google and people can find you with related keywords that you are not using on your web page. What is the rationale behinf including or excludingthis results in search engines index ?

    Read the article

  • Chinese bots in my forum

    - by TdotThomas
    I have a small community forum that doesn't really get posts or any real traffic. The only thing that happens on the regular is bots with Chinese IPs signing up gibberish usernames. Most bots don't make it past the captcha but some do. I try to stay on top of this by banning IPs and ranges of IPs but it doesn't really seem to help. The bots never post anything so what are they doing? Should I be worried? Should I keep banning IPs or is it futile?

    Read the article

  • Why I lose my page rank after 301 redirect?

    - by rajesh.magar
    As we all know Google treats sub-domains as completely separate domains so we have to fight for both, to get ranked in search results. One of my client website was like they having example.com and blog.example.com. So in mind to keep all stuff in one place we redirect blog.example.com to example.com/blog/ But in this case we lost our pagerank and are still wondering where we went wrong or it just takes few more time to showoff. So what is the reason behind this?

    Read the article

  • Form development optimization

    - by Juan
    Like many web developers I do forms all the time. I found myself doing the same all the time: placing input fields, assigning a name to each, ajax the form, then create the PHP which involves to assign a PHP var to each $_REQUEST['var'], escape and validate data, build the html and emailing the results. So I found that 70% of the work is duplicated but I just can't duplicate a page and change the fields. I end up wasting more time reformatting, deleting and adding different fields than creating from scratch. I started planing to program a "list of IDs to html+php" converter in which I'd input all the IDs and this would output the basic html and php. Then I thought: there's got to be thousands of developers that go through this, I'd be reinventing the wheel. So this is my question, I'm trying to find that wheel that somebody must have invented already. I found this: http://www.trirand.com/blog/jqform/ which does more or less what I'm looking for but it's an expensive solution and it has too much functionality for what I'd be using it. Which tools do you use to optimize repetitive task about HTML and PHP?

    Read the article

  • Does the .biz internet top-level domain deserve to be taken seriously?

    - by CrazyJugglerDrummer
    .com is the most preferable TLD most of the time, so many people choosing domain names have to pick between desiredname.something_else or totallydifferent.com. From wikipedia: The biz TLD was created to relieve some of the demand for domain names in the com top-level domain, and to provide an alternative for businesses whose preferred domain name in com had already been registered by another party. .org at least implies an organization, whereas .com is the most generic 'commercial'. But .biz seems like a total duplicate that doesn't seem in widespread use <pun>and deserves to be closed. </pun>. Does .biz have any advantages over .com? Are .biz domains worth registering if .com is already taken?

    Read the article

  • IP blocked because: (smtpauth) Failed SMTP AUTH login from - can someone explain?

    - by gdaniel
    Today I had a few users blocked in our server firewall because of: (smtpauth) Failed SMTP AUTH login from Can someone explain the reason? What does it exactly mean? Could someone be using the our website to access SMTP for spamming purpose? UPDATE: Server info: Centos OS with CPanel and WHM. However no one has access to either. Taking a look at the logs it looks like someone repetitively attempted to login with a known existent user/pass.

    Read the article

  • JavaScript: scroll position (Webkit engine) [migrated]

    - by Julien
    I'm currently trying to use JavaScript to find out how far down the page the user has scrolled; for Firefox 8.0, the keyword is pageYOffset. To say things mechanically: The page has a certain height. In Firefox, the useful object is document.documentElement.scrollHeight. The browser's visible area also has a certain height. In Firefox, the object is window.innerHeight; in IE8, document.documentElement.clientHeight. I need to know where the user is in the page vertically; in other words, how many pixels down the page the user has scrolled. Does Webkit have a DOM object that refers to the current scroll position? Thank you.

    Read the article

< Previous Page | 226 227 228 229 230 231 232 233 234 235 236 237  | Next Page >