Search Results

Search found 2342 results on 94 pages for 'valter minute'.

Page 26/94 | < Previous Page | 22 23 24 25 26 27 28 29 30 31 32 33  | Next Page >

  • Why LLBLGen Pro v3.0 is better than the regular designer for EF v4.0

    As a comment to my blogpost about the 12 minute long video of LLBLGen Pro with Entity Framework v4.0, Burton Roberts asks: Could you list the advantages of using llblgen pro 3 with Entity Framework versus using EF alone? Of course! Now, I know these posts are often classified as 'marketing' but so be it. Why is the LLBLGen Pro v3.0 designer better for Entity Framework v4.0 than the regular Entity Framework designer? Below I'll give an (incomplete) list of features you have at your disposal...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • How to draw a global day night curve

    - by Lumis
    I see many applications which have world-clock map, and I would like to make my own to enhance some of my mobile apps. I wonder if anybody has any knowledge where to start, how to draw a curved shadow representing the dawn and the sunset on the globe. See the example: http://aa.usno.navy.mil/imagery/earth/map?year=2012&month=6&day=19&hour=14&minute=47 I think that this curve goes up and down and creates an artic day/night etc Perhaps there is some acceptable approximation formula without a need to load data for each our and each global parallel and meridian...

    Read the article

  • Long delays in Unity3D substance generation

    - by Josh Buhler
    Currently working on an iOS/Android project in Unity3d, and we're seeing some incredibly long times for generating substances between testing runs. We can run the game, but once we shut down the playback, Unity begins to re-import all off the substances built using Substance Designer. As we've got a lot of these in our game, it's starting to lead to 5 minute delays between testing runs just to test a small change. Any suggestions or parameters we should check that could possibly prevent Unity from needing to regenerate these substances every time? Shouldn't it be caching these things somewhere?

    Read the article

  • Your Supply Chain May Be At Risk for Counterfeiting!

    - by stephen.slade(at)oracle.com
    Competition has driven unscrupulous participants into the supply chain. And your supply chain is exposed along with your customers.   60 Minutes had a terrific segment on fraud in the supply chain. This is a 13 minute clip on our global supply chains and it pertains to us not only as supply chain professionals and participants in supply chain economics, but as consumers and users of the various products that enter our shopping cart.  It's a must see news clip worth sharing   http://www.cbsnews.com/video/watch/?id=7359537n&tag=related;photovideo If you take medicines, you'll want to see this. Dr. Sanjay Gupta participated in this 9 month review of 'bad-drugs' and reports for 60 Minutes on CBS.

    Read the article

  • PostgreSQL data diff

    - by skanatek
    Note: this question is not about syncing database schema/structure Problem In my web application I have a PostgreSQL database server (PGS) and a (separate machine) business logic server (BLS) which regularly (every minute or two) queries 'SELECT ALL' against PGS. The problem is that the 'SELECT ALL' query can easily return 50-200 MB each time. It is obvious that it would be not so good architecture-wise to transfer so much data so frequently over the web. Possible solution What I would like to do is to run some diff tool on PGS and compare the new query with the previous query (all this should be done on PGS). Once the comparison is done I would like to get a dump from PGS and transfer it to BLS. I expect that a diff-based dump would be much, much smaller than the whole 'SELECT ALL' query. Question Is there any data diff tool for PostgreSQL that can do diffs that compare PostgreSQL data between 2 tables or 2 dumps? Note: I would prefer some open-source software tool.

    Read the article

  • Receive the Broadcast program with XML [on hold]

    - by bitmez4
    I have a channel publishing sites and I wanna get into the channel CNN broadcast program.. CNN broadcast the program here: (you can see in source - xml File) http://tvprofil.net/xmltv/data/cnn.info/weekly_cnn.info_tvprofil.net.xml How the data according to the time of withdrawal? For example: Now program: "bitmez's table" next program: "stack's table" in 30 minute Is this possible? UPDATE 1 // -I can take the XML data but to all of XML file- <?php if(!$xml=simplexml_load_file('http://tvprofil.net/xmltv/data/cnn.info/weekly_cnn.info_tvprofil.net.xml')){ trigger_error('XML file -- read error',E_USER_ERROR); } echo 'X-'; foreach($xml as $programme){ echo 'Now: '.$programme->title.' <br/>'; } ?>

    Read the article

  • Searching for Ubuntu video I saw that covered Ubuntu One

    - by Elijah Lynn
    I have been searching for an Ubuntu video I saw 1-2 months ago where they guy was demonstrating Ubuntu One and how he used it to sync local profile stuff across different machines. Has anyone seen this? I forgot exactly what it was about but I know he spent a minute or two on the video if not the whole video. I searched all my web history as well as youtube history but no go. Not sure if we are allowed to ask these questions but here goes.

    Read the article

  • Timeouts in WCF and their default values

      There are a lot of timeouts in WCF. let us summarize it here. Timeouts on binding These are the most well known timeouts. SendTimeout, ReceiveTimeout, OpenTimeout and CloseTimeout. They can be set easily either through config or code on the Binding. The default value for those are 1 minute.  E.g in code Binding binding = new NetTcpBinding(SecurityMode.Transport) { SendTimeout = TimeSpan.FromMinutes(10), ReceiveTimeout = TimeSpan.FromMinutes(10), OpenTimeout...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • What would be the best approach to make revisions of user content?

    - by Kevin Simper
    I have searched and could not find any information about it. What is the best approach to storing revisions? I have a website where the user can write a document which can be fairly long (200-300 lines). How do you determine when to make a revision? Is it not a scalable solution to make a new one whenever the save, because that would be useless to the user when the want to look back, and it would require quite a lot of space. You could use time and say for every 15 minute they are working on it there would be a revision, but that would sometimes be nothing or the whole document have completely changed. I could make a diff from the previous revision, and compare by line and look at how many percent of the lines have been changed. What are other doing revisions?

    Read the article

  • How to Upgrade an existing Customer from OBI10g to 11g: Live Virtual Class

    - by Mike.Hallett(at)Oracle-BI&EPM
    This Live Virtual Class eSeminar on upgrading to Oracle Business Intelligence Enterprise Edition 11g, from OBI 10g, is FREE for Oracle Partners. When : Thursday, January 5, 2012 @ 14.00 CET  / 13.00 UK (120 Minute eSeminar) Where : Goto REGISTER HERE During this session you will learn: OBIEE 11g Infrastructure – What Is Different From OBIEE 10g? Considerations During The Upgrade, Repository Metadata, Presentation Catalog, BI Publisher, BI Security Planning Your Upgrade Optimizing OBI 10g for an 11g Upgrade Copying OBI 10g to New Server Installing OBI 11g on New Server Running Upgrade Assistant & Running OBI 11g Post-Upgrade Steps Testing Upgrading Environment Capacity Planning Guide Q&A Who Should attend? Oracle partners with experience of OBIEE+ 10g BI and EPM developers, architects and implementers Oracle partners with Clients using OBI10g

    Read the article

  • Is It bad for SEO to have internal redirected links? [closed]

    - by Jonas Lindqvist
    I have a large number of pages having similar but not identical content. Example: site.com/dream_dictionary_flying and site.com/dream_interpretation_flying. The problem is that although not being identical, they are sometimes on the edge of being duplicate content. The solution via redirect 301 in htaccess is simple and can be done in a minute, BUT, changing all existing links on the whole site from "/something" to "/something_else" would take ages, it would be thousands of manual changes taking x hundreds of hours. My question is this; is it bad for SEO to have internal links that are redirected, or rather HOW bad is it? For the human user it would not matter at all but from what I have experienced, the search engines don't like it. Is there any rule of thumb here? Please come back with your thoughts and experience on this. Thanks!

    Read the article

  • Windows in StreamInsight: Hopping vs. Snapshot

    - by Roman Schindlauer
    Three weeks ago, we explained the basic concept of windows in StreamInsight: defining sets of events that serve as arguments for set-based operations, like aggregations. Today, we want to discuss the so-called Hopping Windows and compare them with Snapshot Windows. We will compare these two, because they can serve similar purposes with different behaviors; we will discuss the remaining window type, Count Windows, another time. Hopping (and its syntactic-sugar-sister Tumbling) windows are probably the most straightforward windowing concept in StreamInsight. A hopping window is defined by its length, and the offset from one window to the next. They are aligned with some absolute point on the timeline (which can also be given as a parameter to the window) and create sets of events. The diagram below shows an example of a hopping window with length of 1h and hop size (the offset) of 15 minutes, hence creating overlapping windows:   Two aspects in this diagram are important: Since this window is overlapping, an event can fall into more than one windows. If an (interval) event spans a window boundary, its lifetime will be clipped to the window, before it is passed to the set-based operation. That’s the default and currently only available window input policy. (This should only concern you if you are using a time-sensitive user-defined aggregate or operator.) The set-based operation will be applied to each of these sets, yielding a result. This result is: A single scalar value in case of built-in or user-defined aggregates. A subset of the input payloads, in case of the TopK operator. Arbitrary events, when using a user-defined operator. The timestamps of the result are almost always the ones of the windows. Only the user-defined  operator can create new events with timestamps. (However, even these event lifetimes are subject to the window’s output policy, which is currently always to clip to the window end.) Let’s assume we were calculating the sum over some payload field: var result = from window in source.HoppingWindow( TimeSpan.FromHours(1), TimeSpan.FromMinutes(15), HoppingWindowOutputPolicy.ClipToWindowEnd) select new { avg = window.Avg(e => e.Value) }; Now each window is reflected by one result event:   As you can see, the window definition defines the output frequency. No matter how many or few events we got from the input, this hopping window will produce one result every 15 minutes – except for those windows that do not contain any events at all, because StreamInsight window operations are empty-preserving (more about that another time). The “forced” output for every window can become a performance issue if you have a real-time query with many events in a wide group & apply – let me explain: imagine you have a lot of events that you group by and then aggregate within each group – classical streaming pattern. The hopping window produces a result in each group at exactly the same point in time for all groups, since the window boundaries are aligned with the timeline, not with the event timestamps. This means that the query output will become very bursty, delivering the results of all the groups at the same point in time. This becomes especially obvious if the events are long-lasting, spanning multiple windows each, so that the produced result events do not change their value very often. In such a case, a snapshot window can remedy. Snapshot windows are more difficult to explain than hopping windows: they represent those periods in time, when no event changes occur. In other words, if you mark all event start and and times on your timeline, then you are looking at all snapshot window boundaries:   If your events are never overlapping, the snapshot window will not make much sense. It is commonly used together with timestamp modification, which make it a very powerful tool. Or as Allan Mitchell expressed in in a recent tweet: “I used to look at SnapshotWindow() with disdain. Now she is my mistress, the one I turn to in times of trouble and need”. Let’s look at a simple example: I want to compute the average of some value in my events over the last minute. I don’t want this output be produced at fixed intervals, but at soon as it changes (that’s the true event-driven spirit!). The snapshot window will include all currently active event at each point in time, hence we need to extend our original events’ lifetimes into the future: Applying the Snapshot window on these events, it will appear to be “looking back into the past”: If you look at the result produced in this diagram, you can easily prove that, at each point in time, the current event value represents the average of all original input event within the last minute. Here is the LINQ representation of that query, applying the lifetime extension before the snapshot window: var result = from window in source .AlterEventDuration(e => TimeSpan.FromMinutes(1)) .SnapshotWindow(SnapshotWindowOutputPolicy.Clip) select new { avg = window.Avg(e => e.Value) }; With more complex modifications of the event lifetimes you can achieve many more query patterns. For instance “running totals” by keeping the event start times, but snapping their end times to some fixed time, like the end of the day. Each snapshot then “sees” all events that have happened in the respective time period so far. Regards, The StreamInsight Team

    Read the article

  • What are QUICK interview questions for the Microsoft stack development jobs?

    - by Dubmun
    I'm looking for your best "quick answer" technical interview questions. We are a 100% Microsoft shop and do the majority of our development on the ASP.NET web stack in C# and have a custom SOA framework also written in C#. We use a combination of Web Forms, MVC, Web Services, WCF, Entity Framework, SQL Server, TSQL, jQuery, LINQ, and TFS in a SCRUM environment. We are currently on .NET 3.5 with a very near transition to .NET 4.0. Our interviewing process includes a 55 minute interview with two technical people (usually an architect and a senior developer). The two interviewers have to share the time for questions. That isn't enough time for very many true programming problems so I'm looking for more good questions that have quick, yet meaningful, answers. We are mainly interviewing for Senior Dev positions right now but may interview for some Juniors in the future. Please help?

    Read the article

  • How to deal with project managers who micromanage?

    - by entens
    Perhaps I'm just naive, but when I try to decipher the wall of tasks I'm targeted to do over the course of a week, I just can't help but think whoever builds the project schedule needs to get some remedial training on basic project management. For example, I am assigned 13 tasks today, the shortest lasting .13 days (default time metric in Microsoft Project), and the longest lasting .75 days. I can't help but think that it is blatant micromanagement scheduling projects in sub 10 minute intervals. The effects of management are becoming evident in slipped tasks, resource assignment exceeding capacity by a factor of two at some points in time, and spending more time clearing tasks and figuring out what comes next than actually doing work. How can I convince the project manager to create tasks with larger duration and to see the larger picture?

    Read the article

  • Lubuntu 12.04 installation trouble

    - by Evaggelos Tsampazis
    I have i problem! I can't install the Lubuntu 12.04 on my old pc. I only get to the point where I have entered all the needed information (hard drive, username, password etc). When the installation wizard is starting to copy files from the cd to the hard drive, I get after about one minute a blank sreen with only themouse cursor! I can move the mouse cursor but the whole procegure is stoped! Nothing happens after that! My PC: Compaq Pressario 7000 Pentium III 933 MHz 768 MB RAM Any idea how to solve it?? Thanks!!!

    Read the article

  • Temporary dimming of desktop in 12.04

    - by deshmukh
    I am running Ubuntu 12.04 (almost default install regularly updated) Unity interface on ASUS X53U (AMD Brazos Dual Core C60 with 2GB RAM). On launching Thunderbird and Firefox, the application dims and the cursor changes to wait mode. In case of Thunderbird, this is most pronounced with the wait time of up to a minute. Memory status checked with free indicates around 500MB of free memory on such occasions. The OS is stable and I can switch to a different work-space, etc. What could this be? Is this something normal?

    Read the article

  • Consuming JSON stream into AWS Database on the cheap

    - by wjl
    I'm working on a project that needs to consume a JSON stream (approximately 1MB / minute), and parse and insert objects into a database. Amazon's DynamoDB or SimpleDB seem like attractive options for this. Is there a web service that can run a very simple script to eat the data and put it in a database? I could use a worker on Heroku or Elastic Beanstalk, or even pure EC2, but I'd like to find a service that's much cheaper, due to the very low amount of bandwidth and CPU required. (Sorry for the crappy tags. I'm not even sure where to categorize this question.)

    Read the article

  • Sleepyti.me Calculates an Ideal Bedtime Based On Your Morning Schedule

    - by Jason Fitzpatrick
    Sleepyti.me is a web-based sleep calculator that uses average sleep cycle lengths to calculate what time you should go to bed in order to rise at a predetermined well rested and alert. Plug in the time you need to wake up and hit Calculator. Sleepyti.me charts out the 90 minute sleep cycles that will occur over your sleep period and, working backwards, suggests time you should fall asleep in order to wake up in between those cycles in order to increase alertness and have an easier time peeling yourself out of bed. For example, let’s say you need to get up at 7:00 AM. It will suggest you fall asleep at 10:00 PM/11:30 PM/1:00 AM/2:30 AM in order to align your sleep cycles with your proposed rising time. Hit up the link below to take it for a spin. Sleepyti.me How to Own Your Own Website (Even If You Can’t Build One) Pt 3 How to Sync Your Media Across Your Entire House with XBMC How to Own Your Own Website (Even If You Can’t Build One) Pt 2

    Read the article

  • Interview question: How would you implement Google Search?

    - by ripper234
    Supposed you were asked in an interview "How would you implement Google Search?" How would you answer such a question? There might be resources out there that explain how some pieces in Google are implemented (BigTable, MapReduce, PageRank, ...), but that doesn't exactly fit in an interview. What overall architecture would you use, and how would you explain this in a 15-30 minute time span? I would start with explaining how to build a search engine that handles ~ 100k documents, then expand this via sharding to around 50M docs, then perhaps another architectural/technical leap. This is the 20,000 feet view. What I'd like is the details - how you would actually answer that in an interview. Which data structures would you use. What services/machines is your architecture composed of. What would a typical query latency be? What about failover / split brain issues? Etc...

    Read the article

  • Google SMTP relay sending limits

    - by Gavin
    I'm considering using Google Apps for email with my company domain and for sending emails to customers from my website using SMTP. On Google's website it says the following: Limits for registered Google Apps users A registered Google Apps user cannot relay messages to more than 2,000 recipients per day. Limits per domain Per-domain sending limits are determined by the number of users in your Google Apps account. There are two per-domain limits: The maximum number of recipients allowed per domain per day is approximately 130 times the number of users in your Google Apps account. The maximum number of recipients allowed per domain in a 10 minute window is approximately 9 times the number of users in your Google Apps account. Additionally, the maximum number of recipients allowed per domain per day for accounts not yet paid for during the first month of service is 100. If I'm a single user, with a single domain, then does that mean I can only email 130 people a day using SMTP? That limit seems low.

    Read the article

  • Music Rhythm Game Difficulty Question

    - by David Dimalanta
    I have curious question about music rhythm based genre while I'm making a code for the game. Is it really better if I set a random pattern encountered on every music played or there is a specific pattern depending on the music and the difficulty? I have observed the Guitar Hero 3 game for the game console where the difficulty is set on the number of strings used and possible number of combo (e.g. two-string combo). Compared to the Tap Tap Revenge for the Android and iPhone, the difficulty based on the number of BPM (Beat per Minute), meaning, number of targets spawn and must be hit.

    Read the article

  • What framework & technology would you use to make a site like fancy.com [on hold]

    - by adriancdperu
    Im about to start a 3 month process to build something similar to fancy.com. What are your opinions about: best language + framework? (server - side) = I was thinking about LAMP and CakePHP important technological issues to consider when developing? MAU assumption = 1 million. Main features: Registration with Facebook Normal registration Search articles Like articles Post articles Comment on articles Suggest articles to friends via mail, Facebook, Twitter and about 3 or 4 more apis Ranking of articles as a cron job done every minute, many criterias, many rankings Follow users Datamining users to mail everyday with articles they have high probability of liking Operation tools for admins to add articles and close user accounts Mobile focus

    Read the article

  • How to estimate tasks in scrum?

    - by Arian
    Let's say we have a backlog of User Stories, each with an estimated number of Story Points, and now we're doing the Sprint Planning. Now, the Stories should be broken down into tasks and many Scrum resources suggest that each task should be estimated in person-hours. Since all questions have been discussed by the team at this point, estimating a task should not take longer than a minute. However, since a task should not be longer than a day, assuming a three week sprint with 8 developers means 120 tasks, and taking two hours only for estimations seems to be a bit much to me. I know that experienced teams can skip or short-cut task estimations, but let's say we're not at that stage yet. In your experience, how many tasks are there in a sprint* and how long should it take to estimate all of them? (Estimating only half of them doesn't make much sense, does it?) (*) I know that depends on sprint length and team size, so let's assume 8 developers and three weeks.

    Read the article

  • My ubuntuone is broken, what is the problem?

    - by user6962
    I am on 10.10 and u1sdtool seems to be completely broken. I reinstalled it, no change. I can add my PC to the account, but it is added twice(!) every time I tried. So I have my netbook and two times my PC in the account. The netbook with 10.04 has no problems. Below is the error msg I get when attempting to startup Ubuntu One on the command line. desktop:~$ u1sdtool --status Oops, an error ocurred: Traceback (most recent call last): Failure: dbus.exceptions.DBusException: org.freedesktop.DBus.Error.NoReply: Did not receive a reply. Possible causes include: the remote application did not send a reply, the message bus security policy blocked the reply, the reply timeout expired, or the network connection was broken. desktop:~$ Starting it from the Me Menu has the same effect, the HDD will get really busy for a minute and then nothing happens, the client will not start. Nothing in the syslog or anywhere.

    Read the article

  • New Podcast Available: Product Value Chain Management: How Oracle is Taking the Lead on Next Gen Enterprise PLM

    - by Terri Hiskey
    A new podcast on how Oracle is taking the lead in Enterprise PLM with our Product Value Chain solution is now available. In case you're not yet familiar with the concept of Product Value Chain, its an integrated business model powered by Oracle that offers executives the ability to collectively leverage enterprise Agile PLM, Product Data Hub, Enterprise Data Quality and AutoVue Enterprise Visualization and other industry-leading Oracle applications for incremental value. In this quick, 10 minute podcast, you'll hear John Kelley, VP PLM Product Strategy, and Terri Hiskey, Director, PLM Product Marketing, discuss Oracle's vision for next generation enterprise PLM: the Product Value Chain. http://feedproxy.google.com/~r/OracleAppcast/~3/jxAED7ugMEc/11525926_Enterprise_PLM_040612.mp3

    Read the article

< Previous Page | 22 23 24 25 26 27 28 29 30 31 32 33  | Next Page >